Knowledge Distillation Using Output Errors for Self-attention End-to-end Models
暂无分享,去创建一个
Ho-Gyeong Kim | Hwidong Na | Tae Gyoon Kang | Hoshik Lee | Young Sang Choi | Min-Joong Lee | Ji Hyun Lee | Y. S. Choi | Hwidong Na | Min-Joong Lee | Ho-Gyeong Kim | Hoshik Lee | Jihyun Lee
[1] Hisashi Kawai,et al. An Investigation of a Knowledge Distillation Method for CTC Acoustic Models , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[2] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[3] Tara N. Sainath,et al. State-of-the-Art Speech Recognition with Sequence-to-Sequence Models , 2017, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[4] Tara N. Sainath,et al. Minimum Word Error Rate Training for Attention-Based Sequence-to-Sequence Models , 2017, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[5] Chong Wang,et al. Deep Speech 2 : End-to-End Speech Recognition in English and Mandarin , 2015, ICML.
[6] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[7] Shuang Xu,et al. Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[8] Hermann Ney,et al. Improved training of end-to-end attention models for speech recognition , 2018, INTERSPEECH.
[9] Sanjeev Khudanpur,et al. Librispeech: An ASR corpus based on public domain audio books , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[10] Richard Socher,et al. Improving End-to-End Speech Recognition with Policy Learning , 2017, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[11] Kai Yu,et al. Knowledge Distillation for Sequence Model , 2018, INTERSPEECH.
[12] Alexander M. Rush,et al. Sequence-Level Knowledge Distillation , 2016, EMNLP.
[13] Yoshua Bengio,et al. On Using Monolingual Corpora in Neural Machine Translation , 2015, ArXiv.
[14] Quoc V. Le,et al. Listen, Attend and Spell , 2015, ArXiv.
[15] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[16] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[17] Tara N. Sainath,et al. Compression of End-to-End Models , 2018, INTERSPEECH.
[18] Samy Bengio,et al. Tensor2Tensor for Neural Machine Translation , 2018, AMTA.