暂无分享,去创建一个
Matthias Sperber | Jan Niehues | Sebastian Stüker | Alexander H. Waibel | Graham Neubig | A. Waibel | Graham Neubig | J. Niehues | S. Stüker | Matthias Sperber | Sebastian Stüker
[1] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[2] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[3] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[4] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[5] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[6] Jakob Uszkoreit,et al. A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.
[7] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[8] Mari Ostendorf. THE ‘ BEADS-ONA-STRING ’ MODEL OF SPEECH , 1999 .
[9] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[10] Andrew W. Senior,et al. Long short-term memory recurrent neural network architectures for large scale acoustic modeling , 2014, INTERSPEECH.
[11] Quoc V. Le,et al. Listen, Attend and Spell , 2015, ArXiv.
[12] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[13] Paul Deléglise,et al. Enhancing the TED-LIUM Corpus with Selected Data for Language Modeling and More TED Talks , 2014, LREC.
[14] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Mari Ostendorf,et al. Moving beyond the 'beads-on-a-string' model of speech , 1999 .
[16] Ke Li,et al. A Time-Restricted Self-Attention Layer for ASR , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[17] Mirella Lapata,et al. Long Short-Term Memory-Networks for Machine Reading , 2016, EMNLP.
[18] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[19] Matthias Sperber,et al. XNMT: The eXtensible Neural Machine Translation Toolkit , 2018, AMTA.
[20] Yu Zhang,et al. Very deep convolutional networks for end-to-end speech recognition , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[21] Daniel Povey,et al. The Kaldi Speech Recognition Toolkit , 2011 .
[22] Quoc V. Le,et al. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[23] Sungzoon Cho,et al. Distance-based Self-Attention Network for Natural Language Inference , 2017, ArXiv.
[24] Tao Shen,et al. DiSAN: Directional Self-Attention Network for RNN/CNN-free Language Understanding , 2017, AAAI.
[25] David Chiang,et al. Improving Lexical Choice in Neural Machine Translation , 2017, NAACL.
[26] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[27] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.