Content-based methods in peer assessment of open-response questions to grade students as authors and as graders

Massive Open Online Courses (MOOCs) use different types of assignments in order to evaluate student knowledge. Multiple-choice tests are particularly apt given the possibility for automatic assessment of large numbers of assignments. However, certain skills require open responses that cannot be assessed automatically yet their evaluation by instructors or teaching assistants is unfeasible given the large number of students. A potentially effective solution is peer assessment whereby students grade the answers of other students. However, to avoid bias due to inexperience, such grades must be filtered. We describe a factorization approach to grading, as a scalable method capable of dealing with very high volumes of data. Our method is also capable of representing open-response content using a vector space model of the answers. Since reliable peer assessment requires students to make coherent assessments, students can be motivated by their assessments reflecting not only their own answers but also their efforts as graders. The method described is able to tackle both these aspects simultaneously. Finally, for a real-world university setting in Spain, we compared grades obtained by our method and grades awarded by university instructors, with results indicating a notable improvement from using a content-based approach. There was no evidence that instructor grading would have led to more accurate grading outcomes than the assessment produced by our models.

[1]  José Ramón Quevedo,et al.  Feature subset selection for learning preferences: a case study , 2004, ICML.

[2]  Paulo Oliveira,et al.  A system for formative assessment and monitoring of students' progress , 2014, Comput. Educ..

[3]  Amparo Alonso-Betanzos,et al.  Including Content-Based Methods in Peer-Assessment of Open-Response Questions , 2015, 2015 IEEE International Conference on Data Mining Workshop (ICDMW).

[4]  Fakhroddin Noorbehbahani,et al.  The automatic assessment of free text answers using a modified BLEU algorithm , 2011, Comput. Educ..

[5]  Juan José del Coz,et al.  Trait Selection for Assessing Beef Meat Quality Using Non-linear SVM , 2004, NIPS.

[6]  Chun-Yen Chang,et al.  Assessing Creative Problem-solving with Automated Text Grading , 2008 .

[7]  T. Landauer,et al.  Indexing by Latent Semantic Analysis , 1990 .

[8]  Zhenghao Chen,et al.  Tuned Models of Peer Assessment in MOOCs , 2013, EDM.

[9]  Omer Lev,et al.  Impartial Peer Review , 2015, IJCAI.

[10]  Thorsten Joachims,et al.  Bayesian Ordinal Peer Grading , 2015, L@S.

[11]  Roger Blake,et al.  A semantic analysis approach for assessing professionalism using free-form text entered online , 2011, Comput. Hum. Behav..

[12]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[13]  William Barnett,et al.  The modern theory of consumer behavior: Ordinal or cardinal? , 2003 .

[14]  Thorsten Joachims,et al.  Methods for ordinal peer grading , 2014, KDD.

[15]  J. Krosnick,et al.  Survey research. , 1999, Annual review of psychology.

[16]  Gerard Salton,et al.  A vector space model for automatic indexing , 1975, CACM.

[17]  S. García,et al.  An Extension on "Statistical Comparisons of Classifiers over Multiple Data Sets" for all Pairwise Comparisons , 2008 .

[18]  Klaus Obermayer,et al.  Support vector learning for ordinal regression , 1999 .

[19]  Maurizio Magarini,et al.  Weighting peer reviewers , 2014, 2014 Twelfth Annual International Conference on Privacy, Security and Trust.

[20]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[21]  David Maxwell Chickering,et al.  Here or there: preference judgments for relevance , 2008 .

[22]  Justin Cheng,et al.  Peer and self assessment in massive online classes , 2013, ACM Trans. Comput. Hum. Interact..

[23]  Ismael Pascual-Nieto,et al.  Computer-assisted assessment of free-text answers , 2009, The Knowledge Engineering Review.

[24]  Christian Gütl Moving towards a Fully Automatic Knowledge Assessment Tool , 2008, iJET.

[25]  Léon Bottou,et al.  Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.

[26]  P. Sadler,et al.  The Impact of Self- and Peer-Grading on Student Learning , 2006 .

[27]  Amparo Alonso-Betanzos,et al.  A factorization approach to evaluate open-response assignments in MOOCs using preference learning on peer assessments , 2015, Knowl. Based Syst..

[28]  Carlo Strapparava,et al.  Automatic Assessment of Students' Free-Text Answers Underpinned by the Combination of a BLEU-Inspired Algorithm and Latent Semantic Analysis , 2005, FLAIRS Conference.

[29]  Marian Petre,et al.  E-Assessment using Latent Semantic Analysis in the Computer Science Domain: A Pilot Study , 2004 .