Evaluating Human Correction Quality for Machine Translation from Crowdsourcing
暂无分享,去创建一个
[1] Alexander I. Rudnicky,et al. Using the Amazon Mechanical Turk for transcription of spoken language , 2010, 2010 IEEE International Conference on Acoustics, Speech and Signal Processing.
[2] Osamuyimen Stewart,et al. Crowdsourcing participation inequality: a SCOUT model for the enterprise domain , 2010, HCOMP '10.
[3] Dekang Lin,et al. An Information-Theoretic Definition of Similarity , 1998, ICML.
[4] Michael Gamon,et al. Sentence-level MT evaluation without reference translations: beyond language modeling , 2005, EAMT.
[5] Daren C. Brabham. Crowdsourcing as a Model for Problem Solving , 2008 .
[6] Ding Liu,et al. Syntactic Features for Evaluation of Machine Translation , 2005, IEEvaluation@ACL.
[7] Daniel Jurafsky,et al. Parsing to Stanford Dependencies: Trade-offs between Speed and Accuracy , 2010, LREC.
[8] Alex Kulesza,et al. A learning approach to improving sentence-level MT evaluation , 2004 .
[9] David W. Conrath,et al. Semantic Similarity Based on Corpus Statistics and Lexical Taxonomy , 1997, ROCLING/IJCLCLP.
[10] Chris Quirk,et al. Training a Sentence-Level Machine Translation Confidence Measure , 2004, LREC.
[11] Philip Resnik,et al. Using Information Content to Evaluate Semantic Similarity in a Taxonomy , 1995, IJCAI.
[12] Julio Gonzalo,et al. The Contribution of Linguistic Features to Automatic Machine Translation Evaluation , 2009, ACL/IJCNLP.
[13] Chris Callison-Burch,et al. Fast, Cheap, and Creative: Evaluating Translation Quality Using Amazon’s Mechanical Turk , 2009, EMNLP.
[14] Brendan T. O'Connor,et al. Cheap and Fast – But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks , 2008, EMNLP.
[15] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[16] Radu Soricut,et al. TrustRank: Inducing Trust in Automatic Translations via Ranking , 2010, ACL.