CSP: Code-Switching Pre-training for Neural Machine Translation
暂无分享,去创建一个
Zhen Yang | Bojie Hu | Qi Ju | Ambyera Han | Shen Huang | Shen Huang | Zhen Yang | Bojie Hu | Qi Ju | Ambyera Han
[1] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[2] Wei Zhao,et al. Denoising based Sequence-to-Sequence Pre-training for Text Generation , 2019, EMNLP.
[3] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[4] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[5] Wei Chen,et al. Unsupervised Neural Machine Translation with Weight Sharing , 2018 .
[6] Lei Li,et al. Towards Making the Most of BERT in Neural Machine Translation , 2020, AAAI.
[7] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[8] Marjan Ghazvininejad,et al. Multilingual Denoising Pre-training for Neural Machine Translation , 2020, Transactions of the Association for Computational Linguistics.
[9] Shujian Huang,et al. Improving Neural Machine Translation with Pre-trained Representation , 2019, ArXiv.
[10] Tie-Yan Liu,et al. Incorporating BERT into Neural Machine Translation , 2020, ICLR.
[11] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[12] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[13] Martin Wattenberg,et al. Google’s Multilingual Neural Machine Translation System: Enabling Zero-Shot Translation , 2016, TACL.
[14] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[15] Eneko Agirre,et al. Learning bilingual word embeddings with (almost) no bilingual data , 2017, ACL.
[16] Guillaume Lample,et al. Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.
[17] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[18] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[19] Vassilina Nikoulina,et al. On the use of BERT for Neural Machine Translation , 2019, EMNLP.
[20] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[21] Yue Zhang,et al. Code-Switching for Enhancing NMT with Pre-Specified Translation , 2019, NAACL.
[22] Matt Post,et al. Fast Lexically Constrained Decoding with Dynamic Beam Allocation for Neural Machine Translation , 2018, NAACL.
[23] Guillaume Lample,et al. Word Translation Without Parallel Data , 2017, ICLR.
[24] Xuanjing Huang,et al. GlossBERT: BERT for Word Sense Disambiguation with Gloss Knowledge , 2019, EMNLP.
[25] Qun Liu,et al. Lexically Constrained Decoding for Sequence Generation Using Grid Beam Search , 2017, ACL.
[26] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[27] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[28] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[29] Kamel Smaïli,et al. Machine Translation on a Parallel Code-Switched Corpus , 2019, Canadian AI.
[30] Guillaume Lample,et al. Phrase-Based & Neural Unsupervised Machine Translation , 2018, EMNLP.
[31] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[32] Tiejun Zhao,et al. Unsupervised Bilingual Word Embedding Agreement for Unsupervised Neural Machine Translation , 2019, ACL.
[33] Rico Sennrich,et al. Improving Neural Machine Translation Models with Monolingual Data , 2015, ACL.
[34] Shashi Narayan,et al. Leveraging Pre-trained Checkpoints for Sequence Generation Tasks , 2019, Transactions of the Association for Computational Linguistics.
[35] Sergey Edunov,et al. Pre-trained language model representations for language generation , 2019, NAACL.
[36] Eneko Agirre,et al. A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings , 2018, ACL.
[37] Eneko Agirre,et al. Unsupervised Neural Machine Translation , 2017, ICLR.
[38] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.