The LMU Munich Unsupervised Machine Translation System for WMT19
暂无分享,去创建一个
[1] Georgiana Dinu,et al. Improving zero-shot learning by mitigating the hubness problem , 2014, ICLR.
[2] Eneko Agirre,et al. A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings , 2018, ACL.
[3] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[4] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[5] Guillaume Lample,et al. Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.
[6] Eneko Agirre,et al. Unsupervised Neural Machine Translation , 2017, ICLR.
[7] Philipp Koehn,et al. Six Challenges for Neural Machine Translation , 2017, NMT@ACL.
[8] Eneko Agirre,et al. Unsupervised Statistical Machine Translation , 2018, EMNLP.
[9] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[10] Kevin Gimpel,et al. Bridging Nonlinearities and Stochastic Regularizers with Gaussian Error Linear Units , 2016, ArXiv.
[11] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[12] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[13] Guillaume Lample,et al. Word Translation Without Parallel Data , 2017, ICLR.
[14] Alexander M. Fraser,et al. The LMU Munich Unsupervised Machine Translation Systems , 2018, WMT.
[15] Guillaume Lample,et al. Phrase-Based & Neural Unsupervised Machine Translation , 2018, EMNLP.
[16] Zi-Yi Dou,et al. Unsupervised Bilingual Lexicon Induction via Latent Variable Models , 2018, EMNLP.
[17] Matt Post,et al. A Call for Clarity in Reporting BLEU Scores , 2018, WMT.
[18] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[19] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[20] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.