Including Content-Based Methods in Peer-Assessment of Open-Response Questions

Massive Open Online Courses (MOOCs) are attracting the attention of a huge number of students all around the world. These courses include different types of assignments in order to evaluate the student's knowledge. However, these assignments are designed to allow a straightforward automatic evaluation. But, in this way it is not possible to evaluate some skills that would require answering open-response questions. Peer-assessment (the students are asked to assess other assignments), is an effective method to overcome the impossibility of having staff graders for this task. Additionally, students gain a deeper knowledge about the subject of the assignments that they have to read critically. However, the grades given by student-graders must be filtered to avoid bias due to a lack of experience in assessment tasks. There are a number of approaches to do this. In this paper we present a factorization approach that in addition to the grades given by graders is able to incorporate a representation of the contents of the responses given by students using a Vector Space Model of the assignments. So we fill the gap between peer-assessment and content-based methods that use a shallow linguistic processing. The paper includes a report of the results obtained using this approach in a real world dataset collected in 3 universities of Spain, A Coruña, Pablo de Olavide at Sevilla, and Oviedo at Gijón. The scores obtained by the method presented here were compared with those provided by the staff of these universities. We report a considerable improvement whenever we use the content-based approach. In any case, we conclude that there is no evidence that staff grading would have led to more accurate grading outcomes than the assessment produced by our models.

[1]  Zhenghao Chen,et al.  Tuned Models of Peer Assessment in MOOCs , 2013, EDM.

[2]  Nihar B. Shah Some Scaling Laws for MOOC Assessments , 2014 .

[3]  Jorge Dı́ez Peer Assessment in MOOCs Using Preference Learning via Matrix Factorization , 2013 .

[4]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[5]  Kannan Ramchandran,et al.  A Case for Ordinal Peer-evaluation in MOOCs , 2013 .

[6]  Carlo Strapparava,et al.  Automatic Assessment of Students' Free-Text Answers Underpinned by the Combination of a BLEU-Inspired Algorithm and Latent Semantic Analysis , 2005, FLAIRS Conference.

[7]  Marian Petre,et al.  E-Assessment using Latent Semantic Analysis in the Computer Science Domain: A Pilot Study , 2004 .

[8]  V. Aggarwal Principles for using Machine Learning in the Assessment of Open Response Items: Programming Assessment as a Case Study , 2013 .

[9]  Paulo Oliveira,et al.  A system for formative assessment and monitoring of students' progress , 2014, Comput. Educ..

[10]  David Maxwell Chickering,et al.  Here or there: preference judgments for relevance , 2008 .

[11]  T. Landauer,et al.  Indexing by Latent Semantic Analysis , 1990 .

[12]  Christian Gütl Moving towards a Fully Automatic Knowledge Assessment Tool , 2008, iJET.

[13]  Thorsten Joachims,et al.  Methods for ordinal peer grading , 2014, KDD.

[14]  J. Krosnick,et al.  Survey research. , 1999, Annual review of psychology.

[15]  José Ramón Quevedo,et al.  Feature subset selection for learning preferences: a case study , 2004, ICML.

[16]  Fakhroddin Noorbehbahani,et al.  The automatic assessment of free text answers using a modified BLEU algorithm , 2011, Comput. Educ..

[17]  Juan José del Coz,et al.  Trait Selection for Assessing Beef Meat Quality Using Non-linear SVM , 2004, NIPS.

[18]  Roger Blake,et al.  A semantic analysis approach for assessing professionalism using free-form text entered online , 2011, Comput. Hum. Behav..

[19]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[20]  Thorsten Joachims,et al.  Bayesian Ordinal Peer Grading , 2015, L@S.

[21]  Cristina Conati,et al.  AIspace : Interactive Tools for Learning Artificial Intelligence , 2008 .

[22]  William Barnett,et al.  The modern theory of consumer behavior: Ordinal or cardinal? , 2003 .

[23]  P. Sadler,et al.  The Impact of Self- and Peer-Grading on Student Learning , 2006 .

[24]  Justin Cheng,et al.  Peer and self assessment in massive online classes , 2013, ACM Trans. Comput. Hum. Interact..

[25]  Ismael Pascual-Nieto,et al.  Computer-assisted assessment of free-text answers , 2009, The Knowledge Engineering Review.

[26]  Klaus Obermayer,et al.  Support vector learning for ordinal regression , 1999 .

[27]  Amparo Alonso-Betanzos,et al.  A factorization approach to evaluate open-response assignments in MOOCs using preference learning on peer assessments , 2015, Knowl. Based Syst..

[28]  David Maxwell Chickering,et al.  Here or There , 2008, ECIR.

[29]  Gerard Salton,et al.  A vector space model for automatic indexing , 1975, CACM.

[30]  Chun-Yen Chang,et al.  Assessing Creative Problem-solving with Automated Text Grading , 2008 .