Evaluation for morphologically rich language: Russian NLP

RU-EVAL is a biennial event organized in order to estimate the state of the art in Russian NLP resources, methods and toolkits and to compare various methods and principles implemented for Russian. Russian could be treated as an under-resourced language due to the lack of free distributable gold standard corpora for different NLP tasks (each team tried to work out their own standards). Thus, our goal was to work out the uniform basis for comparison of systems based on different theoretical and engineering approaches, to build evaluation resources, to provide a flexible system of evaluation in order to differentiate between non-acceptable and linguistically “admissible” errors. The paper reports on three events devoted to morphological tagging, dependency parsing and anaphora resolution, respectively.

[1]  А Nedoluzhko,et al.  Coreference Chains in Czech, English and Russian: Preliminary Findings. , 2015 .

[2]  Anna Zaretskaya,et al.  Stanford Typed Dependencies: Slavic Languages Application , 2014, PolTAL.

[3]  Toldova S. Ju,et al.  RU-EVAL-2019: EVALUATING ANAPHORA AND COREFERENCE RESOLUTION FOR RUSSIAN , 2014 .

[4]  Antonio Toral,et al.  Evaluation of Natural Language Tools for Italian: EVALITA 2007 , 2008, LREC.

[5]  Maria Vasilyeva,et al.  Evaluating Anaphora and Coreference Resolution for Russian , 2014 .

[6]  Lynette Hirschman,et al.  Appendix F: MUC-7 Coreference Task Definition (version 3.0) , 1998, MUC.

[7]  Olga Lyashevskaya,et al.  RU-EVAL-2012: Evaluating Dependency Parsers for Russian , 2012, COLING.

[8]  Serge Sharoff,et al.  The proper place of men and machines in language technology Processing Russian without any linguistic knowledge , 2011 .

[9]  Aniket Dalal,et al.  Building Feature Rich POS Tagger for Morphologically Rich Languages: Experiences in Hindi , 2022 .

[10]  Joakim Nivre,et al.  MaltEval: an Evaluation and Visualization Tool for Dependency Parsing , 2008, LREC.

[11]  Anastasia Bonch-Osmolovskaya,et al.  Learning Computational Linguistics through NLP Evaluation Events: the experience of Russian evaluation initiative , 2013 .

[12]  Patrick Paroubek,et al.  The GRACE french part-of-speech tagging evaluation task , 1998, LREC.

[13]  Hercules Dalianis,et al.  Automatic training of lemmatization rules that handle morphological changes in pre-, in- and suffixes alike , 2009, ACL.