暂无分享,去创建一个
Lei Xie | Xiaoyu Chen | Xin Lei | Fan Yu | Chao Yang | Zhuoyuan Yao | Xiaoyu Chen | Di Wu | Xiong Wang | Binbin Zhang | Zhendong Peng | Xiangming Wang | X. Lei | Lei Xie | Xiong Wang | Chao Yang | Binbin Zhang | Zhuoyuan Yao | F. Yu | Di Wu | Zhendong Peng | Xiangming Wang
[1] Yoshua Bengio,et al. Attention-Based Models for Speech Recognition , 2015, NIPS.
[2] Shinji Watanabe,et al. Joint CTC-attention based end-to-end speech recognition using multi-task learning , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[3] Jürgen Schmidhuber,et al. Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks , 2006, ICML.
[4] Alex Graves,et al. Sequence Transduction with Recurrent Neural Networks , 2012, ArXiv.
[5] Colin Raffel,et al. Monotonic Chunkwise Attention , 2017, ICLR.
[6] Colin Raffel,et al. Online and Linear-Time Attention by Enforcing Monotonic Alignments , 2017, ICML.
[7] Yafeng Yang,et al. MNN: A Universal and Efficient Inference Engine , 2020, MLSys.
[8] Tara N. Sainath,et al. Universal ASR: Unify and Improve Streaming ASR with Full-context Modeling , 2020, ArXiv.
[9] Yoshua Bengio,et al. End-to-end Continuous Speech Recognition using Attention-based Recurrent NN: First Results , 2014, ArXiv.
[10] Qian Zhang,et al. Transformer Transducer: One Model Unifying Streaming and Non-streaming Speech Recognition , 2020, ArXiv.
[11] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[12] Tara N. Sainath,et al. A Comparison of Sequence-to-Sequence Models for Speech Recognition , 2017, INTERSPEECH.
[13] Chong Wang,et al. Deep Speech 2 : End-to-End Speech Recognition in English and Mandarin , 2015, ICML.
[14] Shinji Watanabe,et al. ESPnet: End-to-End Speech Processing Toolkit , 2018, INTERSPEECH.
[15] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[16] Yu Zhang,et al. Conformer: Convolution-augmented Transformer for Speech Recognition , 2020, INTERSPEECH.
[17] Tara N. Sainath,et al. Two-Pass End-to-End Speech Recognition , 2019, INTERSPEECH.
[18] Hao Zheng,et al. AISHELL-1: An open-source Mandarin speech corpus and a speech recognition baseline , 2017, 2017 20th Conference of the Oriental Chapter of the International Coordinating Committee on Speech Databases and Speech I/O Systems and Assessment (O-COCOSDA).
[19] Quoc V. Le,et al. Listen, Attend and Spell , 2015, ArXiv.
[20] Tatsuya Kawahara,et al. Enhancing Monotonic Multihead Attention for Streaming ASR , 2020, INTERSPEECH.
[21] Lei Xie,et al. Unified Streaming and Non-streaming Two-pass End-to-end Model for Speech Recognition , 2020, ArXiv.
[22] Quoc V. Le,et al. SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition , 2019, INTERSPEECH.
[23] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.