Correct-and-Memorize: Learning to Translate from Interactive Revisions

State-of-the-art machine translation models are still not on par with human translators. Previous work takes human interactions into the neural machine translation process to obtain improved results in target languages. However, not all model-translation errors are equal -- some are critical while others are minor. In the meanwhile, the same translation mistakes occur repeatedly in a similar context. To solve both issues, we propose CAMIT, a novel method for translating in an interactive environment. Our proposed method works with critical revision instructions, therefore allows human to correct arbitrary words in model-translated sentences. In addition, CAMIT learns from and softly memorizes revision actions based on the context, alleviating the issue of repeating mistakes. Experiments in both ideal and real interactive translation settings demonstrate that our proposed \method enhances machine translation results significantly while requires fewer revision instructions from human compared to previous methods.

[1]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[2]  Zhiming Chen,et al.  Neural Post-Editing Based on Quality Estimation , 2017, WMT.

[3]  Gonzalo Iglesias,et al.  Neural Machine Translation Decoding with Terminology Constraints , 2018, NAACL.

[4]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[5]  Kevin Knight,et al.  A Syntax-based Statistical Translation Model , 2001, ACL.

[6]  George F. Foster,et al.  TransType: a Computer-Aided Translation Typing System , 2000 .

[7]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[8]  Rico Sennrich,et al.  Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.

[9]  Dianhai Yu,et al.  Multi-Task Learning for Multiple Language Translation , 2015, ACL.

[10]  David Chiang,et al.  Hierarchical Phrase-Based Translation , 2007, CL.

[11]  Shujian Huang,et al.  PRIMT: A Pick-Revise Framework for Interactive Machine Translation , 2016, NAACL.

[12]  Qun Liu,et al.  Lexically Constrained Decoding for Sequence Generation Using Grid Beam Search , 2017, ACL.

[13]  Yang Liu,et al.  Learning to Remember Translation History with a Continuous Cache , 2017, TACL.

[14]  Shujian Huang,et al.  Combining Character and Word Information in Neural Machine Translation Using a Multi-Level Attention , 2018, NAACL.

[15]  Francisco Casacuberta,et al.  Active Learning for Interactive Neural Machine Translation of Data Streams , 2018, CoNLL.

[16]  John DeNero,et al.  Models and Inference for Prefix-Constrained Machine Translation , 2016, ACL.

[17]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[18]  Lemao Liu,et al.  Agreement on Target-bidirectional Neural Machine Translation , 2016, NAACL.

[19]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[20]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[21]  Roland Kuhn,et al.  Rule-Based Translation with Statistical Phrase-Based Post-Editing , 2007, WMT@ACL.

[22]  Mauro Cettolo,et al.  WIT3: Web Inventory of Transcribed and Translated Talks , 2012, EAMT.

[23]  Quoc V. Le,et al.  Multi-task Sequence to Sequence Learning , 2015, ICLR.

[24]  Philipp Koehn,et al.  Neural Interactive Translation Prediction , 2016, AMTA.

[25]  Francisco Casacuberta,et al.  Interactive neural machine translation , 2017, Comput. Speech Lang..

[26]  Matt Post,et al.  Fast Lexically Constrained Decoding with Dynamic Beam Allocation for Neural Machine Translation , 2018, NAACL.

[27]  Francisco Casacuberta,et al.  Online Learning for Neural Machine Translation Post-editing , 2017, ArXiv.

[28]  Zaixiang Zheng,et al.  Learning to Discriminate Noises for Incorporating External Information in Neural Machine Translation , 2018, ArXiv.

[29]  Hang Li,et al.  “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .

[30]  Jiancheng Lv,et al.  BFGAN: Backward and Forward Generative Adversarial Networks for Lexically Constrained Sentence Generation , 2018, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[31]  Daniel Marcu,et al.  Statistical Phrase-Based Translation , 2003, NAACL.

[32]  Germán Sanchis-Trilles,et al.  Interactive translation prediction versus conventional post-editing in practice: a study with the CasMaCat  workbench , 2014, Machine Translation.

[33]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[34]  Rui Yan,et al.  Sequence to Backward and Forward Sequences: A Content-Introducing Approach to Generative Short-Text Conversation , 2016, COLING.

[35]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[36]  Jason Weston,et al.  Memory Networks , 2014, ICLR.