Developing an automated writing placement system for ESL learners

ABSTRACT There are quite a few challenges in the development of an automated writing placement model for non-native English learners, among them the fact that exams that encompass the full range of language proficiency exhibited at different stages of learning are hard to design. However, acquisition of appropriate training data that are relevant to the task at hand is essential in the development of the model. Using the Cambridge Learner Corpus writing scores, which have been subsequently benchmarked to Common European Framework of Reference for Languages (CEFR) levels, we conceptualize the task as a supervised machine learning problem, and primarily focus on developing a generic writing model. Such an approach facilitates the modeling of truly consistent, internal marking criteria regardless of the prompt delivered, which has the additional advantage of requiring smaller dataset sizes and not necessarily requiring re-training or tuning for new tasks. The system is developed to predict someone’s proficiency level on the CEFR scale, which allows learners to point to a specific standard of achievement. We furthermore integrate our model into Cambridge English Write & ImproveTM—a freely available, cloud-based tool that automatically provides diagnostic feedback to non-native English language learners at different levels of granularity—and examine its use.

[1]  Semire Dikli,et al.  An Overview of Automated Scoring of Essays. , 2006 .

[2]  Salvatore Valenti,et al.  An Overview of Current Research on Automated Essay Grading , 2003, J. Inf. Technol. Educ..

[3]  Sandra Kübler,et al.  Predicting Learner Levels for Online Exercises of Hebrew , 2012, BEA@NAACL-HLT.

[4]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .

[5]  Helen Yannakoudakis,et al.  A New Dataset and Method for Automatically Grading ESOL Texts , 2011, ACL.

[6]  Ben Hamner,et al.  Contrasting state-of-the-art automated scoring of essays: analysis , 2012 .

[7]  D Nicholls,et al.  The Cambridge Learner Corpus-Error coding and analysis , 1999 .

[8]  Elena Volodina,et al.  Classification of Language Proficiency Levels in Swedish Learners’ Texts , 2016 .

[9]  Jianfeng Gao,et al.  Using Statistical Techniques and Web Search to Correct ESL Errors , 2013 .

[10]  Martin Chodorow,et al.  CriterionSM Online Essay Evaluation: An Application for Automated Evaluation of Student Essays , 2003, IAAI.

[11]  David M. Williamson A Framework for Implementing Automated Scoring , 2009 .

[12]  Torsten Zesch,et al.  Predicting proficiency levels in learner writings by transferring a linguistic complexity model from expert-written coursebooks , 2016, COLING.

[13]  Ying-Jian Wang,et al.  Exploring the impact of using automated writing evaluation in English as a foreign language university students' writing , 2013 .

[14]  Jacob Cohen,et al.  Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit. , 1968 .

[15]  Helen Yannakoudakis,et al.  Classifying intermediate learner English: A data-driven approach to learner corpora , 2013 .

[16]  Florencio Lopez-de-Silanes,et al.  Overview of Current Research , 2005 .

[17]  Erkki Sutinen,et al.  Semi-Automatic Evaluation Features in Computer-assisted Essay Assessment , 2004, CATE.

[18]  Ted Briscoe,et al.  The Second Release of the RASP System , 2006, ACL.

[19]  Graeme Hirst,et al.  Robust, Lexicalized Native Language Identification , 2012, COLING.

[20]  Martin Chodorow,et al.  Automated Essay Evaluation: The Criterion Online Writing Service , 2004, AI Mag..

[21]  Helen Yannakoudakis,et al.  Automating Second Language Acquisition Research: Integrating Information Visualisation and Machine Learning , 2012, EACL 2012.

[22]  Jill Burstein,et al.  AUTOMATED ESSAY SCORING WITH E‐RATER® V.2.0 , 2004 .

[23]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[24]  E. B. Page,et al.  The use of the computer in analyzing student essays , 1968 .

[25]  D. H I G G I N S,et al.  Identifying off-topic student essays without topic-specific training data † , 2005 .

[26]  David Alfter,et al.  Classification of Swedish learner essays by CEFR levels , 2016 .

[27]  Helen Yannakoudakis,et al.  Developing and testing a self-assessment and tutoring system , 2013, BEA@NAACL-HLT.

[28]  Jill Burstein,et al.  The E-rater® scoring engine: Automated essay scoring with natural language processing. , 2003 .

[29]  Jason S. Chang,et al.  WriteAhead2: Mining Lexical Grammar Patterns for Assisted Writing , 2015, HLT-NAACL.

[30]  Ted Briscoe,et al.  An introduction to tag sequence grammars and the RASP system parser , 2006 .

[31]  Daniel Marcu,et al.  Finding the WRITE Stuff: Automatic Identification of Discourse Structure in Student Essays , 2003, IEEE Intell. Syst..

[32]  P. S. Gingrich,et al.  The writer's workbench: Computer aids for text analysis , 1982 .

[33]  Ted Briscoe,et al.  Automated assessment of ESOL free text examinations , 2010 .

[34]  Silvia Bernardini,et al.  Introducing and evaluating ukWaC , a very large web-derived corpus of English , 2008 .

[35]  吉島 茂,et al.  文化と言語の多様性の中のCommon European Framework of Reference for Languages: Learning, teaching, assessment (CEFR)--それは基準か? (第10回明海大学大学院応用言語学研究科セミナー 講演) , 2008 .

[36]  Rada Mihalcea,et al.  Using Word Semantics To Assist English as a Second Language Learners , 2015, HLT-NAACL.

[37]  Sowmya Vajjala,et al.  Automatic CEFR Level Prediction for Estonian Learner Text , 2014 .

[38]  Akiko Aizawa,et al.  CroVeWA: Crosslingual Vector-Based Writing Assistance , 2015, NAACL.

[39]  Helen Yannakoudakis,et al.  Unsupervised Modeling of Topical Relevance in L2 Learner Text , 2016, BEA@NAACL-HLT.