Quality Estimation for Language Output Applications

Quality Estimation (QE) of language output applications is a research area that has been attracting significant attention. The goal of QE is to estimate the quality of language output applications without the need of human references. Instead, machine learning algorithms are used to build supervised models based on a few labelled training instances. Such models are able to generalise over unseen data and thus QE is a robust method applicable to scenarios where human input is not available or possible. One such a scenario where QE is particularly appealing is that of Machine Translation, where a score for predicted quality can help decide whether or not a translation is useful (e.g. for post-editing) or reliable (e.g. for gisting). Other potential applications within Natural Language Processing (NLP) include Text Summarisation and Text Simplification. In this tutorial we present the task of QE and its application in NLP, focusing on Machine Translation. We also introduce QuEst++, a toolkit for QE that encompasses feature extraction and machine learning, and propose a practical activity to extend this toolkit in various ways.

[1]  Radu Soricut,et al.  TrustRank: Inducing Trust in Automatic Translations via Ranking , 2010, ACL.

[2]  Lucia Specia,et al.  Exploiting Objective Annotations for Minimising Translation Post-editing Effort , 2011, EAMT.

[3]  Yifan He,et al.  Bridging SMT and TM with Translation Recommendation , 2010, ACL.

[4]  Lucia Specia,et al.  Phrase Level Segmentation and Labelling of Machine Translation Errors , 2016, LREC.

[5]  Lucia Specia,et al.  Multi-level Translation Quality Prediction with QuEst++ , 2015, ACL.

[6]  Karin M. Verspoor,et al.  Findings of the 2016 Conference on Machine Translation , 2016, WMT.

[7]  Benjamin Lecouteux,et al.  An Open Source Toolkit for Word-level Confidence Estimation in Machine Translation , 2015 .

[8]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[9]  Lucia Specia,et al.  QuEst - A translation quality estimation framework , 2013, ACL.

[10]  Lucia Specia,et al.  MARMOT: A Toolkit for Translation Quality Estimation at the Word Level , 2016, LREC.

[11]  Lucia Specia,et al.  Document-level translation quality estimation: exploring discourse and pseudo-references , 2014, EAMT.

[12]  Lucia Specia,et al.  Phrase-level Quality Estimation for Machine Translation , 2015 .

[13]  Josef van Genabith,et al.  Searching for Context: a Study on Document-Level Labels for Translation Quality Estimation , 2015, EAMT.

[14]  Lucia Specia,et al.  Quality estimation for translation selection , 2014, EAMT.

[15]  Lluís Màrquez i Villodre,et al.  A Graphical Interface for MT Evaluation and Error Analysis , 2012, ACL.