Investigating Catastrophic Forgetting During Continual Training for Neural Machine Translation
暂无分享,去创建一个
[1] Yang Feng,et al. Improving Domain Adaptation Translation with Domain Invariant and Specific Information , 2019, NAACL.
[2] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[3] Huda Khayrallah,et al. Overcoming Catastrophic Forgetting During Domain Adaptation of Neural Machine Translation , 2019, NAACL.
[4] Yang Feng,et al. Bridging the Gap between Training and Inference for Neural Machine Translation , 2019, ACL.
[5] Tuo Zhao,et al. Multi-Domain Neural Machine Translation with Word-Level Adaptive Layer-wise Domain Mixing , 2019, ACL.
[6] Yonatan Belinkov,et al. Identifying and Controlling Important Neurons in Neural Machine Translation , 2018, ICLR.
[7] John DeNero,et al. Compact Personalized Models for Neural Machine Translation , 2018, EMNLP.
[8] Deniz Yuret,et al. Why Neural Translations are the Right Length , 2016, EMNLP.
[9] Huda Khayrallah,et al. Regularized Training Objective for Continued Training for Domain Adaptation in Neural Machine Translation , 2018, NMT@ACL.
[10] Chenhui Chu,et al. An Empirical Comparison of Domain Adaptation Methods for Neural Machine Translation , 2017, ACL.
[11] Rico Sennrich,et al. Regularization techniques for fine-tuning in neural machine translation , 2017, EMNLP.
[12] Graham Neubig,et al. When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation? , 2018, NAACL.
[13] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[14] Fedor Moiseev,et al. Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned , 2019, ACL.
[15] Dakwale,et al. Fine-Tuning for Neural Machine Translation with Limited Degradation across In- and Out-of-Domain Data , 2017, MTSUMMIT.
[16] Chenhui Chu,et al. An Empirical Comparison of Simple Domain Adaptation Methods for Neural Machine Translation , 2017, ArXiv.
[17] Markus Freitag,et al. Fast Domain Adaptation for Neural Machine Translation , 2016, ArXiv.
[18] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.
[19] Jie Zhou,et al. Token-level Adaptive Training for Neural Machine Translation , 2020, EMNLP.
[20] Liang Tian,et al. UM-Corpus: A Large English-Chinese Parallel Corpus for Statistical Machine Translation , 2014, LREC.
[21] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[22] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[23] Timo Aila,et al. Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.
[24] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[25] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[26] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[27] Huda Khayrallah,et al. Freezing Subnetworks to Analyze Domain Adaptation in Neural Machine Translation , 2018, WMT.
[28] Yang Liu,et al. Visualizing and Understanding Neural Machine Translation , 2017, ACL.
[29] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[30] Pierre Alquier,et al. A Theoretical Analysis of Catastrophic Forgetting through the NTK Overlap Matrix , 2020, AISTATS.
[31] Xilin Chen,et al. Greedy Search with Probabilistic N-gram Matching for Neural Machine Translation , 2018, EMNLP.
[32] Christopher D. Manning,et al. Stanford Neural Machine Translation Systems for Spoken Language Domains , 2015, IWSLT.
[33] Rico Sennrich,et al. On Exposure Bias, Hallucination and Domain Shift in Neural Machine Translation , 2020, ACL.