Pipeline Signed Japanese Translation Focusing on a Post-positional Particle Complement and Conjugation in a Low-resource Setting

Because sign language is a visual language, the translation of it into spoken language is typically performed through an intermediate representation called gloss notation. In sign language, function words, such as particles and determiners, are not explicitly expressed, and there is little or no concept of morphological inflection in sign language. Therefore, gloss notation does not include such linguistic constructs. Because of these factors, we argue that sign language translation is effectively processed by taking advantage of the similarities and differences between sign language and its spoken counterpart. We thus propose a pipeline translation method that clearly focuses on the difference between spoken Japanese and signed Japanese written in gloss notation. Specifically, our method first uses statistical machine translation (SMT) to map glosses to corresponding spoken language words. We then use three transformer-based seq2seq models trained using a large out-ofdomain monolingual Japanese corpus to complement postpositional particles and estimate conjugations for the verbs, adjectives, and auxiliary verbs in the first translation. We apply the seq2seq models in sequence until the translation converges. Our experimental results show that the proposed method performs robustly on the low-resource corpus and is +4.4/+4.9 points above the SMT baseline for BLEU-3/4.

[1]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[2]  Myle Ott,et al.  fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.

[3]  Alon Lavie,et al.  METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments , 2005, IEEvaluation@ACL.

[4]  Mohamed Jemni,et al.  English-ASL Gloss Parallel Corpus 2012: ASLG-PC12 , 2012 .

[5]  Rico Sennrich,et al.  Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.

[6]  John O. Isenhath The Linguistics of American Sign Language , 1990 .

[7]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[8]  Hermann Ney,et al.  Neural Sign Language Translation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[9]  Yuji Matsumoto,et al.  Towards Automatic Error Type Classification of Japanese Language Learners’ Writings , 2013, PACLIC.

[10]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[11]  Philipp Koehn,et al.  Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.

[12]  Oscar Koller,et al.  Sign Language Transformers: Joint End-to-End Sign Language Recognition and Translation , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Jesse Read,et al.  Better Sign Language Translation with STMC-Transformer , 2020, COLING.

[14]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[15]  Masaru Takeuchi,et al.  An Annotated Japanese Sign Language Corpus , 2002, LREC.

[16]  Helen F. Hastie,et al.  Transfer Learning for British Sign Language Modelling , 2018, VarDial@COLING 2018.

[17]  Kugatsu Sadamitsu,et al.  Grammar Error Correction Using Pseudo-Error Sentences and Domain Adaptation , 2012, ACL.

[18]  Andy Way,et al.  The ATIS Sign Language Corpus , 2008, LREC.

[19]  Hiroyuki Shindo,et al.  Automatic Error Correction on Japanese Functional Expressions Using Character-based Neural Machine Translation , 2018, PACLIC.

[20]  Rico Sennrich,et al.  On Exposure Bias, Hallucination and Domain Shift in Neural Machine Translation , 2020, ACL.