Neural Machine Translation with Deep Attention
暂无分享,去创建一个
Jinsong Su | Biao Zhang | Deyi Xiong | Biao Zhang | Deyi Xiong | Jinsong Su
[1] Kenneth Heafield,et al. N-gram Counts and Language Models from the Common Crawl , 2014, LREC.
[2] Hermann Ney,et al. A Systematic Comparison of Various Statistical Alignment Models , 2003, CL.
[3] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[4] Alexander J. Smola,et al. Stacked Attention Networks for Image Question Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Yang Liu,et al. Coverage-based Neural Machine Translation , 2016, ArXiv.
[6] Shan Wu,et al. Variational Recurrent Neural Machine Translation , 2018, AAAI.
[7] Philipp Koehn,et al. Statistical Significance Tests for Machine Translation Evaluation , 2004, EMNLP.
[8] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[9] Deyi Xiong,et al. Recurrent Neural Machine Translation , 2016, ArXiv.
[10] Marcello Federico,et al. Deep Neural Machine Translation with Weakly-Recurrent Units , 2018, EAMT.
[11] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[12] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[13] Yang Liu,et al. Modeling Coverage for Neural Machine Translation , 2016, ACL.
[14] Qun Liu,et al. Memory-enhanced Decoder for Neural Machine Translation , 2016, EMNLP.
[15] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[16] Yoshua Bengio,et al. On Using Very Large Target Vocabulary for Neural Machine Translation , 2014, ACL.
[17] Yang Liu,et al. Towards Robust Neural Machine Translation , 2018, ACL.
[18] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[19] Rongrong Ji,et al. Asynchronous Bidirectional Decoding for Neural Machine Translation , 2018, AAAI.
[20] Yang Liu,et al. Contrastive Unsupervised Word Alignment with Non-Local Features , 2014, AAAI.
[21] Jürgen Schmidhuber,et al. Training Very Deep Networks , 2015, NIPS.
[22] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[24] Yang Liu,et al. Minimum Risk Training for Neural Machine Translation , 2015, ACL.
[25] Qun Liu,et al. Deep Neural Machine Translation with Linear Associative Unit , 2017, ACL.
[26] Wei Xu,et al. Deep Recurrent Models with Fast-Forward Connections for Neural Machine Translation , 2016, TACL.
[27] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[28] Deyi Xiong,et al. Accelerating Neural Transformer via an Average Attention Network , 2018, ACL.
[29] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[30] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[31] Deyi Xiong,et al. A GRU-Gated Attention Model for Neural Machine Translation , 2017, ArXiv.
[32] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.