Incentives for Truthful Peer Grading

Peer grading systems work well only if users have incentives to grade truthfully. An example of non-truthful grading, that we observed in classrooms, consists in students assigning the maximum grade to all submissions. With a naive grading scheme, such as averaging the assigned grades, all students would receive the maximum grade. In this paper, we develop three grading schemes that provide incentives for truthful peer grading. In the first scheme, the instructor grades a fraction p of the submissions, and penalizes students whose grade deviates from the instructor grade. We provide lower bounds on p to ensure truthfulness, and conclude that these schemes work only for moderate class sizes, up to a few hundred students. To overcome this limitation, we propose a hierarchical extension of this supervised scheme, and we show that it can handle classes of any size with bounded (and little) instructor work, and is therefore applicable to Massive Open Online Courses (MOOCs). Finally, we propose unsupervised incentive schemes, in which the student incentive is based on statistical properties of the grade distribution, without any grading required by the instructor. We show that the proposed unsupervised schemes provide incentives to truthful grading, at the price of being possibly unfair to individual students.

[1]  R. Zeckhauser,et al.  Efficiency Despite Mutually Payoff-Relevant Private Information: The Finite Case , 1990 .

[2]  D. Prelec A Bayesian Truth Serum for Subjective Data , 2004, Science.

[3]  Eric Horvitz,et al.  Incentives for truthful reporting in crowdsourcing , 2012, AAMAS.

[4]  Nir Ailon,et al.  Aggregation of Partial Rankings, p-Ratings and Top-m Lists , 2007, SODA '07.

[5]  Boi Faltings,et al.  Mechanisms for Making Crowds Truthful , 2014, J. Artif. Intell. Res..

[6]  Arpita Ghosh,et al.  Game Theory and Incentives in Human Computation Systems , 2013, Handbook of Human Computation.

[7]  Noga Alon,et al.  Sum of us: strategyproof selection from the selectors , 2009, TARK XIII.

[8]  Boi Faltings,et al.  Minimum payments that reward honest reputation feedback , 2006, EC '06.

[9]  Robert T. Clemen,et al.  Incentive contrats and strictly proper scoring rules , 2002 .

[10]  A. H. Murphy,et al.  “Good” Probability Assessors , 1968 .

[11]  Kate Larson,et al.  Inducing Honest Reporting Without Observing Outcomes: An Application to the Peer-Review Process , 2013, 1309.3197.

[12]  Thorsten Joachims,et al.  Methods for ordinal peer grading , 2014, KDD.

[13]  Ariel Rubinstein,et al.  A Course in Game Theory , 1995 .

[14]  Anirban Dasgupta,et al.  Crowdsourced judgement elicitation with endogenous proficiency , 2013, WWW.

[15]  Boi Faltings,et al.  Enforcing Truthful Strategies in Incentive Compatible Reputation Mechanisms , 2005, WINE.

[16]  Paul Resnick,et al.  Eliciting Informative Feedback: The Peer-Prediction Method , 2005, Manag. Sci..

[17]  Richard M. Karp,et al.  Reducibility Among Combinatorial Problems , 1972, 50 Years of Integer Programming.

[18]  Omer Lev,et al.  Impartial Peer Review , 2015, IJCAI.

[19]  Luca de Alfaro,et al.  CrowdGrader: a tool for crowdsourcing the evaluation of homework assignments , 2014, SIGCSE.

[20]  Toby Walsh,et al.  The PeerRank Method for Peer Assessment , 2014, ECAI.

[21]  Moni Naor,et al.  Rank aggregation methods for the Web , 2001, WWW '01.