UCSYNLP-Lab Machine Translation Systems for WAT 2018

This paper describes the UCSYNLP-Lab submission to WAT 2019 for Myanmar-English translation tasks in both direction. We have used the neural machine translation systems with attention model and utilized the UCSY-corpus and ALT corpus. In NMT with attention model, we use the word segmentation level as well as syllable segmentation level. Especially, we made the UCSY-corpus to be cleaned in WAT 2019. Therefore, the UCSY corpus for WAT 2019 is not identical to those used in WAT 2018. Experiments show that the translation systems can produce the substantial improvements.

[1]  Sadao Kurohashi,et al.  Kyoto University Participation to WAT 2017 , 2017, WAT@IJCNLP.

[2]  Khin Mar Soe,et al.  Large Scale Myanmar to English Neural Machine Translation System , 2018, 2018 IEEE 7th Global Conference on Consumer Electronics (GCCE).

[3]  Khin Mar Soe,et al.  Myanmar Phrases Translation Model with Morphological Analysis for Statistical Myanmar to English Translation System , 2011, PACLIC.

[4]  Rico Sennrich,et al.  Improving Neural Machine Translation Models with Monolingual Data , 2015, ACL.

[5]  Masaaki Nagata,et al.  NTT Neural Machine Translation Systems at WAT 2017 , 2019, WAT@IJCNLP.

[6]  Eiichiro Sumita,et al.  A Study of Statistical Machine Translation Methods for Under Resourced Languages , 2016, SLTU.

[7]  David Chiang,et al.  Hierarchical Phrase-Based Translation , 2007, CL.

[8]  Alexander M. Rush,et al.  OpenNMT: Open-Source Toolkit for Neural Machine Translation , 2017, ACL.

[9]  Matt Post,et al.  We start by defining the recurrent architecture as implemented in S OCKEYE , following , 2018 .

[10]  Kenneth Heafield,et al.  KenLM: Faster and Smaller Language Model Queries , 2011, WMT@EMNLP.

[11]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[12]  Hermann Ney,et al.  Improved Statistical Alignment Models , 2000, ACL.

[13]  Christoph Tillmann,et al.  A Unigram Orientation Model for Statistical Machine Translation , 2004, NAACL.

[14]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[15]  F ChenStanley,et al.  An Empirical Study of Smoothing Techniques for Language Modeling , 1996, ACL.

[16]  Stanley F. Chen,et al.  An Empirical Study of Smoothing Techniques for Language Modeling , 1996, ACL.

[17]  Haizhou Li,et al.  Adequacy–Fluency Metrics: Evaluating MT in the Continuous Space Model Framework , 2015, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[18]  Franz Josef Och,et al.  Minimum Error Rate Training in Statistical Machine Translation , 2003, ACL.