暂无分享,去创建一个
Tie-Yan Liu | Houqiang Li | Wengang Zhou | Yingce Xia | Jinhua Zhu | Di He | Lijun Wu | Tao Qin
[1] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[2] Gregory Shakhnarovich,et al. FractalNet: Ultra-Deep Neural Networks without Residuals , 2016, ICLR.
[3] Rico Sennrich,et al. Edinburgh Neural Machine Translation Systems for WMT 16 , 2016, WMT.
[4] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[5] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[6] Marcin Junczys-Dowmunt,et al. MS-UEdin Submission to the WMT2018 APE Shared Task: Dual-Source Transformer for Automatic Post-Editing , 2018, WMT.
[7] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.
[8] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[9] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[10] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[11] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[12] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[13] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[14] Guillaume Lample,et al. Phrase-Based & Neural Unsupervised Machine Translation , 2018, EMNLP.
[15] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[16] James Henderson,et al. Document-Level Neural Machine Translation with Hierarchical Attention Networks , 2018, EMNLP.
[17] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[18] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[19] Myle Ott,et al. Scaling Neural Machine Translation , 2018, WMT.
[20] Marc'Aurelio Ranzato,et al. Classical Structured Prediction Losses for Sequence to Sequence Learning , 2017, NAACL.
[21] Jason Weston,et al. Contextualized Role Interaction for Neural Machine Translation , 2018 .
[22] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[23] Quoc V. Le,et al. The Evolved Transformer , 2019, ICML.
[24] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[25] Rico Sennrich,et al. Improving Neural Machine Translation Models with Monolingual Data , 2015, ACL.
[26] Pascal Vincent,et al. The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training , 2009, AISTATS.
[27] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[28] Lijun Wu,et al. Learning to Teach with Dynamic Loss Functions , 2018, NeurIPS.
[29] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[30] Yann Dauphin,et al. Pay Less Attention with Lightweight and Dynamic Convolutions , 2019, ICLR.
[31] Zhe Gan,et al. Distilling the Knowledge of BERT for Text Generation , 2019, ArXiv.
[32] Alexander M. Rush,et al. Latent Alignment and Variational Attention , 2018, NeurIPS.
[33] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[34] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[35] Di He,et al. Tied Transformers: Neural Machine Translation with Shared Encoder and Decoder , 2019, AAAI.
[36] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[37] Marcello Federico,et al. Report on the 11th IWSLT evaluation campaign , 2014, IWSLT.
[38] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[39] Lei Li,et al. Towards Making the Most of BERT in Neural Machine Translation , 2020, AAAI.
[40] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[41] Marta R. Costa-jussà,et al. Findings of the 2019 Conference on Machine Translation (WMT19) , 2019, WMT.