Deep Recurrent Models with Fast-Forward Connections for Neural Machine Translation
暂无分享,去创建一个
Wei Xu | Peng Li | Jie Zhou | Ying Cao | Xuguang Wang | W. Xu | Jie Zhou | Ying Cao | Xuguang Wang | Peng Li
[1] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[2] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[3] Mikel L. Forcada,et al. Recursive Hetero-associative Memories for Translation , 1997, IWANN.
[4] Joan Cabestany,et al. Biological and Artificial Computation: From Neuroscience to Technology , 1997, Lecture Notes in Computer Science.
[5] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[6] Daniel Marcu,et al. Statistical Phrase-Based Translation , 2003, NAACL.
[7] Ben Taskar,et al. Alignment by Agreement , 2006, NAACL.
[8] J. Schmidhuber,et al. A Novel Connectionist System for Unconstrained Handwriting Recognition , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[9] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[10] Yoshua Bengio,et al. Practical Recommendations for Gradient-Based Training of Deep Architectures , 2012, Neural Networks: Tricks of the Trade.
[11] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[12] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.
[13] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[14] Kenneth Heafield,et al. N-gram Counts and Language Models from the Common Crawl , 2014, LREC.
[15] Nadir Durrani,et al. Edinburgh’s Phrase-based Machine Translation Systems for WMT-14 , 2014, WMT@ACL.
[16] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[17] Yoshua Bengio,et al. On Using Very Large Target Vocabulary for Neural Machine Translation , 2014, ACL.
[18] Quoc V. Le,et al. Addressing the Rare Word Problem in Neural Machine Translation , 2014, ACL.
[19] Wei Xu,et al. Deep Captioning with Multimodal Recurrent Neural Networks (m-RNN) , 2014, ICLR.
[20] Quoc V. Le,et al. A Neural Conversational Model , 2015, ArXiv.
[21] Yang Yu,et al. Empirical Study on Deep Learning Models for QA , 2015 .
[22] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[23] Ekaterina Vylomova,et al. Depth-Gated LSTM , 2015, ArXiv.
[24] Wei Xu,et al. End-to-end learning of semantic role labeling using recurrent neural networks , 2015, ACL.
[25] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[26] Yang Yu,et al. Empirical Study on Deep Learning Models for QA , 2015, 1510.07526.
[27] Jürgen Schmidhuber,et al. Highway Networks , 2015, ArXiv.
[28] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[30] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Alex Graves,et al. Grid Long Short-Term Memory , 2015, ICLR.