Linking open-source code commits and MOOC grades to evaluate massive online open peer review

Massive Open Online Courses (MOOCs) have been used by students as a low-cost and low-touch educational credential in a variety of fields. Understanding the grading mechanisms behind these course assignments is important for evaluating MOOC credentials. A common approach to grading free-response assignments is massive scale peer-review, specially used for assignments that are not easy to grade programmatically. It is difficult to assess these approaches since the responses typically require human evaluation. Here we link data from public code repositories on GitHub and 1 ar X iv :2 10 4. 12 55 5v 1 [ cs .C Y ] 1 5 A pr 2 02 1 course grades for a large massive-online open course to study the dynamics of massive scale peer review. This has important implications for understanding the dynamics of difficult to grade assignments. Since the research was not hypothesis driven, we described the results in an exploratory framework. We find three distinct clusters of repeated peer-review submissions and use these clusters to study how grades change in response to changes in code submissions. Our exploration also leads to an important observation that massive scale peer-review scores are highly variable, increase, on average, with repeated submissions, and changes in scores are not closely tied to the code changes that form the basis for the re-submissions. Keywords— data analysis, MOOC education, GitHub code commits, peer grading

[1]  Li Yuan,et al.  MOOCs and open education: Implications for higher education , 2013 .

[2]  D. Bonett,et al.  Sample size requirements for estimating pearson, kendall and spearman correlations , 2000 .

[3]  Adrian E. Raftery,et al.  How Many Clusters? Which Clustering Method? Answers Via Model-Based Cluster Analysis , 1998, Comput. J..

[4]  John D. Blischak,et al.  A Quick Introduction to Version Control with Git and GitHub , 2016, PLoS Comput. Biol..

[5]  J. Cavanaugh,et al.  The Bayesian information criterion: background, derivation, and applications , 2012 .

[6]  Sarah E. M. Meek,et al.  Is peer review an appropriate form of assessment in a MOOC? Student participation and performance in formative peer review , 2017 .

[7]  Guoliang Li,et al.  Truth Inference in Crowdsourcing: Is the Problem Solved? , 2017, Proc. VLDB Endow..

[8]  Douglas A. Reynolds,et al.  Gaussian Mixture Models , 2018, Encyclopedia of Biometrics.

[9]  Shivakumar Vaithyanathan,et al.  Model-Based Hierarchical Clustering , 2000, UAI.

[10]  Sebastián Ventura,et al.  Educational data science in massive open online courses , 2016, WIREs Data Mining Knowl. Discov..

[11]  Adrian E. Raftery,et al.  mclust Version 4 for R : Normal Mixture Modeling for Model-Based Clustering , Classification , and Density Estimation , 2012 .

[12]  Luca Scrucca,et al.  mclust 5: Clustering, Classification and Density Estimation Using Gaussian Finite Mixture Models , 2016, R J..

[13]  D. B. Owen,et al.  The Power of Student's t-test , 1965 .

[14]  Malik Beshir Malik,et al.  Applied Linear Regression , 2005, Technometrics.

[15]  Doug Clow,et al.  MOOCs and the funnel of participation , 2013, LAK '13.

[16]  Hoi K. Suen,et al.  Peer assessment for massive open online courses (MOOCs) , 2014 .