LAYERED: Metric for Machine Translation Evaluation
暂无分享,去创建一个
[1] Andy Way,et al. Evaluating machine translation with LFG dependencies , 2007, Machine Translation.
[2] Daniel Jurafsky,et al. Robust Machine Translation Evaluation with Entailment Features , 2009, ACL.
[3] Alon Lavie,et al. METEOR: An Automatic Metric for MT Evaluation with High Levels of Correlation with Human Judgments , 2007, WMT@ACL.
[4] Ondrej Bojar,et al. Results of the WMT14 Metrics Shared Task , 2013 .
[5] Pushpak Bhattacharyya,et al. Some Issues in Automatic Evaluation of English-Hindi MT: More Blues for BLEU , 2006 .
[6] Philipp Koehn,et al. Findings of the 2013 Workshop on Statistical Machine Translation , 2013, WMT@ACL.
[7] Adam Lopez,et al. Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies , 2011 .
[8] Alexandra Birch,et al. Reordering Metrics for MT , 2011, ACL.
[9] George R. Doddington,et al. Automatic Evaluation of Machine Translation Quality Using N-gram Co-Occurrence Statistics , 2002 .
[10] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[11] Lluís Màrquez i Villodre,et al. Linguistic measures for automatic machine translation evaluation , 2010, Machine Translation.
[12] Christopher D. Manning,et al. Generating Typed Dependency Parses from Phrase Structure Parses , 2006, LREC.
[13] Matthew G. Snover,et al. A Study of Translation Edit Rate with Targeted Human Annotation , 2006, AMTA.
[14] Ying Zhang,et al. Interpreting BLEU/NIST Scores: How Much Improvement do We Need to Have a Better System? , 2004, LREC.
[15] Alexandra Birch,et al. Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies , 2011 .
[16] Ding Liu,et al. Syntactic Features for Evaluation of Machine Translation , 2005, IEEvaluation@ACL.
[17] Ralph Weischedel,et al. A STUDY OF TRANSLATION ERROR RATE WITH TARGETED HUMAN ANNOTATION , 2005 .