Recurrent deep neural networks for robust speech recognition C Weng, D Yu, S Watanabe, BHF Juang 2014 IEEE International Conference on Acoustics, Speech and Signal …, 2014 | 156 | 2014 |
Durian: Duration informed attention network for multimodal synthesis C Yu, H Lu, N Hu, M Yu, C Weng, K Xu, P Liu, D Tuo, S Kang, G Lei, D Su, ... arXiv preprint arXiv:1909.01700, 2019 | 150 | 2019 |
Deep neural networks for single-channel multi-talker speech recognition C Weng, D Yu, ML Seltzer, J Droppo IEEE/ACM Transactions on Audio, Speech, and Language Processing 23 (10 …, 2015 | 103 | 2015 |
Component fusion: Learning replaceable language model component for end-to-end speech recognition system C Shan, C Weng, G Wang, D Su, M Luo, D Yu, L Xie ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 83 | 2019 |
Gigaspeech: An evolving, multi-domain asr corpus with 10,000 hours of transcribed audio G Chen, S Chai, G Wang, J Du, WQ Zhang, C Weng, D Su, D Povey, ... arXiv preprint arXiv:2106.06909, 2021 | 77 | 2021 |
Replay and synthetic speech detection with res2net architecture X Li, N Li, C Weng, X Liu, D Su, D Yu, H Meng ICASSP 2021-2021 IEEE international conference on acoustics, speech and …, 2021 | 74 | 2021 |
Investigating end-to-end speech recognition for mandarin-english code-switching C Shan, C Weng, G Wang, D Su, M Luo, D Yu, L Xie ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 70 | 2019 |
Past review, current progress, and challenges ahead on the cocktail party problem Y Qian, C Weng, X Chang, S Wang, D Yu Frontiers of Information Technology & Electronic Engineering 19, 40-63, 2018 | 69 | 2018 |
Improving Attention Based Sequence-to-Sequence Models for End-to-End English Conversational Speech Recognition. C Weng, J Cui, G Wang, J Wang, C Yu, D Su, D Yu Interspeech, 761-765, 2018 | 59 | 2018 |
Mixed speech recognition D Yu, C Weng, ML Seltzer, J Droppo US Patent 9,390,712, 2016 | 51 | 2016 |
Single-channel mixed speech recognition using deep neural networks C Weng, D Yu, ML Seltzer, J Droppo 2014 IEEE International Conference on Acoustics, Speech and Signal …, 2014 | 45 | 2014 |
Self-supervised text-independent speaker verification using prototypical momentum contrastive learning W Xia, C Zhang, C Weng, M Yu, D Yu ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and …, 2021 | 40 | 2021 |
Deep learning based multi-source localization with source splitting and its effectiveness in multi-talker speech recognition AS Subramanian, C Weng, S Watanabe, M Yu, D Yu Computer Speech & Language 75, 101360, 2022 | 38 | 2022 |
Joint training of complex ratio mask based beamformer and acoustic model for noise robust asr Y Xu, C Weng, L Hui, J Liu, M Yu, D Su, D Yu ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 35 | 2019 |
Minimum bayes risk training of rnn-transducer for end-to-end speech recognition C Weng, C Yu, J Cui, C Zhang, D Yu arXiv preprint arXiv:1911.12487, 2019 | 34 | 2019 |
Far-field location guided target speech extraction using end-to-end speech recognition objectives AS Subramanian, C Weng, M Yu, SX Zhang, Y Xu, S Watanabe, D Yu ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020 | 32 | 2020 |
Durian-sc: Duration informed attention network based singing voice conversion system L Zhang, C Yu, H Lu, C Weng, C Zhang, Y Wu, X Xie, Z Li, D Yu arXiv preprint arXiv:2008.03009, 2020 | 31 | 2020 |
Neural spatio-temporal beamformer for target speech separation Y Xu, M Yu, SX Zhang, L Chen, C Weng, J Liu, D Yu arXiv preprint arXiv:2005.03889, 2020 | 30 | 2020 |
Non-autoregressive transformer asr with ctc-enhanced decoder input X Song, Z Wu, Y Huang, C Weng, D Su, H Meng ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and …, 2021 | 28 | 2021 |
Pitchnet: Unsupervised singing voice conversion with pitch adversarial network C Deng, C Yu, H Lu, C Weng, D Yu ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020 | 28 | 2020 |