暂无分享,去创建一个
[1] Ying Zhang,et al. Batch normalized recurrent neural networks , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[2] Hermann Ney,et al. Empirical Investigation of Optimization Algorithms in Neural Machine Translation , 2017, Prague Bull. Math. Linguistics.
[3] Victor O. K. Li,et al. Non-Autoregressive Neural Machine Translation , 2017, ICLR.
[4] Jianxin Wu,et al. Minimal gated unit for recurrent neural networks , 2016, International Journal of Automation and Computing.
[5] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[6] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[7] Philipp Koehn,et al. Findings of the 2017 Conference on Machine Translation (WMT17) , 2017, WMT.
[8] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[9] Yu Zhang,et al. Training RNNs as Fast as CNNs , 2017, EMNLP 2018.
[10] Alexander M. Rush,et al. OpenNMT: Open-Source Toolkit for Neural Machine Translation , 2017, ACL.
[11] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[12] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[13] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[14] Mauro Cettolo,et al. WIT3: Web Inventory of Transcribed and Translated Talks , 2012, EAMT.
[15] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[16] Arianna Bisazza,et al. Neural versus Phrase-Based Machine Translation Quality: a Case Study , 2016, EMNLP.
[17] Rico Sennrich,et al. Nematus: a Toolkit for Neural Machine Translation , 2017, EACL.
[18] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[19] Hava T. Siegelmann,et al. On the Computational Power of Neural Nets , 1995, J. Comput. Syst. Sci..
[20] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[21] Alex Graves,et al. Neural Machine Translation in Linear Time , 2016, ArXiv.
[22] Lior Wolf,et al. Using the Output Embedding to Improve Language Models , 2016, EACL.
[23] Regina Barzilay,et al. Deriving Neural Architectures from Sequence and Graph Kernels , 2017, ICML.
[24] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[25] Ankur Bapna,et al. The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation , 2018, ACL.
[26] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[27] Rico Sennrich,et al. Deep architectures for Neural Machine Translation , 2017, WMT.
[28] Razvan Pascanu,et al. How to Construct Deep Recurrent Neural Networks , 2013, ICLR.
[29] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[30] Christopher D. Manning,et al. Stanford Neural Machine Translation Systems for Spoken Language Domains , 2015, IWSLT.
[31] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[32] Rico Sennrich,et al. Edinburgh Neural Machine Translation Systems for WMT 16 , 2016, WMT.
[33] Muhammad Ghifary,et al. Strongly-Typed Recurrent Neural Networks , 2016, ICML.
[34] Karin M. Verspoor,et al. Findings of the 2016 Conference on Machine Translation , 2016, WMT.
[35] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[36] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[37] Yoshua Bengio,et al. Light Gated Recurrent Units for Speech Recognition , 2018, IEEE Transactions on Emerging Topics in Computational Intelligence.