Acceptability of machine-translated content: A multi-language evaluation by translators and end-users

As machine translation (MT) continues to be used increasingly in the translation industry, there is a corresponding increase in the need to understand MT quality and, in particular, its impact on end-users. To date, little work has been carried out to investigate the acceptability of MT output among end-users and, ultimately, how acceptable they find it. This article reports on research conducted to address that gap. End-users of instructional content machine-translated from English into German, Simplified Chinese and Japanese were engaged in a usability experiment. Part of this experiment involved giving feedback on the acceptability of raw machine-translated content and lightly post-edited (PE) versions of the same content. In addition, a quality review was carried out in collaboration with an industry partner and experienced translation quality reviewers. The translation quality-assessment (TQA) results from translators reflect the usability and satisfaction results by end-users insofar as the implementation of light PE both increased the usability and acceptability of the PE instructions and led to satisfaction being reported. Nonetheless, the raw MT content also received good scores, especially for terminology, country standards and spelling.

[1]  Sheila Castilho,et al.  Does post-editing increase usability? A study with Brazilian Portuguese as target language , 2014, EAMT.

[2]  Lucia Specia,et al.  Exploiting Objective Annotations for Minimising Translation Post-editing Effort , 2011, EAMT.

[3]  Joanna Drugan,et al.  Quality In Professional Translation: Assessment and Improvement , 2013 .

[4]  H. Bernard Research Methods in Anthropology: Qualitative and Quantitative Approaches , 1988 .

[5]  Noam Chomsky,et al.  वाक्यविन्यास का सैद्धान्तिक पक्ष = Aspects of the theory of syntax , 1965 .

[6]  Ana Guerberof Arenas Correlations between productivity and quality when post-editing in a professional context , 2014, Machine Translation.

[7]  Michael Carl,et al.  Post-editing machine translation: A usability test for professional translation settings , 2015 .

[8]  H. Isahara,et al.  Evaluation method for determining groups of users who find MT “useful” , 2001, MTSUMMIT.

[9]  Ilse Depraetere,et al.  What counts as useful advice in a university post-editing training context? Report on a case , 2010, EAMT.

[10]  R. Beaugrande,et al.  Introduction to text linguistics , 1981 .

[11]  Sara Stymne,et al.  Eye Tracking as a Tool for Machine Translation Error Analysis , 2012, LREC.

[12]  Sharon O'Brien,et al.  Analysing Post-Editing Performance: Correlations with Years of Translation Experience , 2010, EAMT.

[13]  Robert J. Hartsuiker,et al.  The impact of machine translation error types on post-editing effort indicators , 2015, MTSUMMIT.

[14]  Geoffrey S. Koby,et al.  Defining Translation Quality , 2014 .

[15]  Sheila Castilho,et al.  Content profiling and translation scenarios , 2016 .

[16]  Jakob Nielsen,et al.  Usability engineering , 1997, The Computer Science and Engineering Handbook.

[17]  Sharon O'Brien,et al.  Correlations of perceived post-editing effort with measurements of actual effort , 2015, Machine Translation.

[18]  Maarit Koponen,et al.  Comparing human perceptions of post-editing effort with post-editing operations , 2012, WMT@NAACL-HLT.

[19]  D. Cramer,et al.  Introduction to Statistics in Psychology , 2005 .

[20]  Masaru Tomita,et al.  Evaluation of MT Systems by TOEFL , 1993, TMI.

[21]  Andrei Popescu-Belis,et al.  Principles of Context-Based Machine Translation Evaluation , 2002, Machine Translation.

[22]  Jody Byrne,et al.  Technical Translation: Usability Strategies for Translating Technical Documentation , 2006 .

[23]  Stephen Doherty,et al.  Eye tracking as an MT evaluation technique , 2010, Machine Translation.

[24]  Lucia Specia,et al.  Assessing the Post-Editing Effort for Automatic and Semi-Automatic Translations of DVD Subtitles , 2011, RANLP.

[25]  Douglas A. Reynolds,et al.  Measuring human readability of machine generated text: three case studies in speech recognition and machine translation , 2005, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005..

[26]  Sharon O'Brien,et al.  Assessing the Usability of Raw Machine Translated Output: A User-Centered Study Using Eye Tracking , 2014, Int. J. Hum. Comput. Interact..

[27]  François Masselot,et al.  A Productivity Test of Statistical Machine Translation Post-Editing in a Typical Localisation Context , 2010, Prague Bull. Math. Linguistics.

[28]  A. Burchardt,et al.  Multidimensional Quality Metrics (MQM): A Framework for Declaring and Describing Translation Quality Metrics , 2014 .