CLASSIFICATION OF MT-OUTPUT USING HYBRID MT-EVALUATION METRICS FOR POST-EDITING AND THEIR COMPARATIVE STUDY
暂无分享,去创建一个
[1] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[2] Alon Lavie,et al. METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments , 2005, IEEvaluation@ACL.
[3] Alon Lavie,et al. METEOR-NEXT and the METEOR Paraphrase Tables: Improved Evaluation Support for Five Target Languages , 2010, WMT@ACL.
[4] Matthew G. Snover,et al. A Study of Translation Edit Rate with Targeted Human Annotation , 2006, AMTA.
[5] Alexander H. Waibel,et al. The CMU statistical machine translation system for IWSLT 2005 , 2005, IWSLT.
[6] Ian R. Lane,et al. The UKA/CMU statistical machine translation system for IWSLT 2006 , 2006, IWSLT.
[7] Hermann Ney,et al. Accelerated DP based search for statistical translation , 1997, EUROSPEECH.
[8] François Masselot,et al. A Productivity Test of Statistical Machine Translation Post-Editing in a Typical Localisation Context , 2010, Prague Bull. Math. Linguistics.
[9] Kenneth Ward Church,et al. Good applications for crummy machine translation , 1993, Machine Translation.
[10] Kevin Knight,et al. Automated Postediting of Documents , 1994, AAAI.
[11] Ralph Weischedel,et al. A STUDY OF TRANSLATION ERROR RATE WITH TARGETED HUMAN ANNOTATION , 2005 .
[12] Michel Simard,et al. Statistical Phrase-Based Post-Editing , 2007, NAACL.
[13] Georges Van Slype,et al. Evaluation du systeme de traduction automatique SYSTRAN anglais-Francais, version 1978, de la Commission des Communautes Europeennes (Evaluation of the English-French SYSTRAN Machine Translation System, 19 version, of the Commission of European Communities). , 1979 .