暂无分享,去创建一个
Deyi Xiong | Mingxuan Wang | Lei Li | Tao Wang | Chengqi Zhao | Hang Li
[1] Jacob Eisenstein,et al. AdvAug: Robust Adversarial Augmentation for Neural Machine Translation , 2020, ACL.
[2] Josef van Genabith,et al. How Robust Are Character-Based Word Embeddings in Tagging and MT Against Wrod Scramlbing or Randdm Nouse? , 2017, AMTA.
[3] A. Waibel,et al. Toward Robust Neural Machine Translation for Noisy Input Sequences , 2017, IWSLT.
[4] Graham Neubig,et al. Improving Robustness of Neural Machine Translation with Multi-task Learning , 2019, WMT.
[5] Yonatan Belinkov,et al. Synthetic and Natural Noise Both Break Neural Machine Translation , 2017, ICLR.
[6] Taro Watanabe,et al. Denoising Neural Machine Translation Training with Trusted Data and Online Data Selection , 2018, WMT.
[7] Omer Levy,et al. Training on Synthetic Noise Improves Robustness to Natural Noise in Machine Translation , 2019, EMNLP.
[8] Keith Stevens,et al. Effective Parallel Corpus Mining using Bilingual Sentence Embeddings , 2018, WMT.
[9] Matt Post,et al. A Call for Clarity in Reporting BLEU Scores , 2018, WMT.
[10] Zhongjun He,et al. Robust Neural Machine Translation with Joint Textual and Phonetic Embedding , 2018, ACL.
[11] Ciprian Chelba,et al. Dynamically Composing Domain-Data Selection with Clean-Data Selection by “Co-Curricular Learning” for Neural Machine Translation , 2019, ACL.
[12] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[13] Changhan Wang,et al. Levenshtein Transformer , 2019, NeurIPS.
[14] Christof Monz,et al. Dynamic Data Selection for Neural Machine Translation , 2017, EMNLP.
[15] Huda Khayrallah,et al. Findings of the WMT 2018 Shared Task on Parallel Corpus Filtering , 2018, WMT.
[16] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[17] Akshay Chaturvedi,et al. Exploring the Robustness of NMT Systems to Nonsensical Inputs , 2019 .
[18] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[19] Graham Neubig,et al. On Evaluation of Adversarial Perturbations for Sequence-to-Sequence Models , 2019, NAACL.
[20] Yong Cheng,et al. Robust Neural Machine Translation with Doubly Adversarial Inputs , 2019, ACL.
[21] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[22] Masao Utiyama,et al. Dynamic Sentence Sampling for Efficient Training of Neural Machine Translation , 2018, ACL.
[23] Proceedings of the Fifth Conference on Machine Translation, WMT@EMNLP 2020, Online, November 19-20, 2020 , 2020, WMT@EMNLP.
[24] Dejing Dou,et al. On Adversarial Examples for Character-Level Neural Machine Translation , 2018, COLING.
[25] Holger Schwenk,et al. Filtering and Mining Parallel Data in a Joint Multilingual Space , 2018, ACL.
[26] Graham Neubig,et al. Target Conditioned Sampling: Optimizing Data Selection for Multilingual Neural Machine Translation , 2019, ACL.
[27] Philipp Koehn,et al. Zipporah: a Fast and Scalable Data Cleaning System for Noisy Web-Crawled Parallel Corpora , 2017, EMNLP.
[28] Graham Neubig,et al. Improving Robustness of Machine Translation with Synthetic Noise , 2019, NAACL.
[29] Sameer Singh,et al. Generating Natural Adversarial Examples , 2017, ICLR.
[30] Lei Li,et al. Autocorrect in the Process of Translation — Multi-task Learning Improves Dialogue Machine Translation , 2021, NAACL.