Principles for Assessing Adaptive Online Courses

Adaptive online courses are designed to automatically customize material for different users, typically based on data captured during the course. Assessing the quality of these adaptive courses, however, can be difficult. Traditional assessment methods for (machine) learning algorithms, such as comparison against a ground truth, are often unavailable due to education’s unique goal of affecting both internal user knowledge, which cannot be directly measured, as well as external, measurable performance. Traditional metrics for education like quiz scores, on the other hand, do not necessarily capture the adaptive course’s ability to present the right material to different users. In this work, we present a mathematical framework for developing scalable, efficiently computable metrics for these courses that can be used by instructors to gauge the efficacy of the adaptation and their course content. Our metric framework takes as input a set of quantities describing user activities in the course, and balances definitions of user consistency and overall efficacy as inferred by the quantity distributions. We support the metric definitions by comparing the results of a comprehensive statistical analysis with a sample metric evaluation on a dataset of roughly 5,000 users from an online chess platform. In doing so, we find that our metrics yield important insights about the course that are embedded in the larger statistical analysis, as well as additional insights into student drop-off rates.

[1]  Maliha S. Nash,et al.  Handbook of Parametric and Nonparametric Statistical Procedures , 2001, Technometrics.

[2]  René F. Kizilcec,et al.  Motivation as a Lens to Understand Online Learners , 2015, ACM Trans. Comput. Hum. Interact..

[3]  Jie Xu,et al.  Predicting Grades , 2015, IEEE Transactions on Signal Processing.

[4]  Sean M. McNee,et al.  Being accurate is not enough: how accuracy metrics have hurt recommender systems , 2006, CHI Extended Abstracts.

[5]  Richard G. Baraniuk,et al.  Time-varying learning and content analytics via sparse factor analysis , 2013, KDD.

[6]  David E. Pritchard,et al.  Studying Learning in the Worldwide Classroom Research into edX's First MOOC. , 2013 .

[7]  Katrien Verbert,et al.  HCI for Recommender Systems: the Past, the Present and the Future , 2016, RecSys.

[8]  Li Yujian,et al.  A Normalized Levenshtein Distance Metric , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  J. Meigs,et al.  WHO Technical Report , 1954, The Yale Journal of Biology and Medicine.

[10]  Michael Jahrer,et al.  Collaborative Filtering Applied to Educational Data Mining , 2010 .

[11]  Guy Shani,et al.  A Survey of Accuracy Evaluation Metrics of Recommendation Tasks , 2009, J. Mach. Learn. Res..

[12]  Mung Chiang,et al.  Behavior in social learning networks: Early detection for online short-courses , 2017, IEEE INFOCOM 2017 - IEEE Conference on Computer Communications.

[13]  Gwo-Jen Hwang,et al.  Development of an Adaptive Learning System with Multiple Perspectives based on Students? Learning Styles and Cognitive Styles , 2013, J. Educ. Technol. Soc..

[14]  Krzysztof Z. Gajos,et al.  Understanding in-video dropouts and interaction peaks in online lecture videos Citation , 2014 .

[15]  Zhenming Liu,et al.  Learning about Social Learning in MOOCs: From Statistical Analysis to Generative Model , 2013, IEEE Transactions on Learning Technologies.

[16]  Sangtae Ha,et al.  Individualization for Education at Scale: MIIC Design and Preliminary Evaluation , 2015, IEEE Transactions on Learning Technologies.

[17]  David Hinkley,et al.  Bootstrap Methods: Another Look at the Jackknife , 2008 .

[18]  Sebastián Ventura,et al.  Predicting students' final performance from participation in on-line discussion forums , 2013, Comput. Educ..

[19]  Alan Said,et al.  Comparative recommender system evaluation: benchmarking recommendation frameworks , 2014, RecSys '14.

[20]  Qian Zhang,et al.  Modeling and Predicting Learning Behavior in MOOCs , 2016, WSDM.

[21]  N. Heffernan,et al.  Using HMMs and bagged decision trees to leverage rich features of user and skill from an intelligent tutoring system dataset , 2010 .

[22]  Carla Limongelli,et al.  Adaptive Learning with the LS-Plan System: A Field Evaluation , 2009, IEEE Transactions on Learning Technologies.

[23]  H. Vincent Poor,et al.  Mining MOOC Clickstreams: Video-Watching Behavior vs. In-Video Quiz Performance , 2016, IEEE Transactions on Signal Processing.