Multimodal Learning Analytics and Education Data Mining: using computational technologies to measure complex learning tasks

New high-frequency multimodal data collection technologies and machine learning analysis techniques could offer new insights into learning, especially when students have the opportunity to generate unique, personalized artifacts, such as computer programs, robots, and solutions engineering challenges. To date most of the work on learning analytics and educational data mining has been focused on online courses and cognitive tutors, both of which provide a high degree of structure to the tasks, and are restricted to interactions that occur in front of a computer screen. In this paper, we argue that multimodal learning analytics can offer new insights into students’ learning trajectories in more complex and open-ended learning environments. We present several examples of this work and its educational application.

[1]  Kenneth R. Koedinger,et al.  Adapting handwriting recognition for applications in algebra learning , 2007, Emme '07.

[2]  Dor Abrahamson,et al.  The mathematical imagery trainer: from embodied interaction to conceptual learning , 2011, CHI.

[3]  Marcelo Worsley,et al.  Towards the development of multimodal action based assessment , 2013, LAK '13.

[4]  Bertrand Schneider,et al.  Unraveling Students' Interaction Around a Tangible Interface Using Gesture Recognition , 2014, EDM.

[5]  Simon Harper,et al.  Using galvanic skin response measures to identify areas of frustration for older web 2.0 users , 2010, W4A.

[6]  Johanna D. Moore,et al.  Using Natural Language Processing to Analyze Tutorial Dialogue Corpora Across Domains Modalities , 2009, AIED.

[7]  Tanja Schultz,et al.  Vision-based handwriting recognition for unrestricted text input in mid-air , 2012, ICMI '12.

[8]  J. Bhattacharya,et al.  High-Learners Present Larger Mid-Frontal Theta Power and Connectivity in Response to Incorrect Performance Feedback , 2013, The Journal of Neuroscience.

[9]  Zachary A. Pardos,et al.  Affective states and state tests: investigating how affect throughout the school year predicts end of year learning outcomes , 2013, LAK '13.

[10]  Benjamin D. Jee,et al.  Drawing on Experience: Use of Sketching to Evaluate Knowledge of Spatial Scientific Concepts , 2009 .

[11]  Marcelo Worsley,et al.  Using learning analytics to study cognitive disequilibrium in a complex learning environment , 2015, LAK.

[12]  Bertrand Schneider,et al.  Real-time mutual gaze perception enhances collaborative learning and collaboration quality , 2013, International Journal of Computer-Supported Collaborative Learning.

[13]  Patrick Jermann,et al.  Collaborative Gaze Footprints: Correlates of Interaction Quality , 2011, CSCL.

[14]  Pierre Dillenbourg,et al.  Sleepers' lag - study on motion and attention , 2014, LAK.

[15]  Mubarak Shah,et al.  Actions sketch: a novel action representation , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[16]  Milena K. Nigam,et al.  The Equivalence of Learning Paths in Early Science Instruction: Effects of Direct Instruction and Discovery Learning , 2022 .

[17]  Victor R. Lee,et al.  Framing in cognitive clinical interviews about intuitive science knowledge: Dynamic student understandings of the discourse interaction , 2012 .

[18]  Jelena Obradović,et al.  The construct of psychophysiological reactivity: Statistical and psychometric issues , 2013 .

[19]  Christine Alvarado,et al.  A Framework for Multi-Domain Sketch Recognition , 2002 .

[20]  Juhan Nam,et al.  Multimodal Deep Learning , 2011, ICML.

[21]  Janet C. Read,et al.  A study of the usability of handwriting recognition for text entry by children , 2007, Interact. Comput..

[22]  Kenneth D. Forbus,et al.  Using Quantitative Information to Improve Analogical Matching Between Sketches , 2012, IAAI.

[23]  Marcelo Worsley,et al.  Deciphering the Practices and Affordances of Different Reasoning Strategies through Multimodal Learning Analytics , 2014, MLA@ICMI.

[24]  Günther Palm,et al.  A generic framework for the inference of user states in human computer interaction , 2012, Journal on Multimodal User Interfaces.

[25]  Osmar R. Zaïane,et al.  Web Usage Mining for a Better Web-Based Learning Environment , 2001 .

[26]  Seymour Papert,et al.  Mindstorms: Children, Computers, and Powerful Ideas , 1981 .

[27]  Ronald H. Stevens,et al.  EEG-Related Changes in Cognitive Workload, Engagement and Distraction as Students Acquire Problem Solving Skills , 2007, User Modeling.

[28]  Omar AlZoubi,et al.  Affect Detection from Multichannel Physiology during Learning Sessions with AutoTutor , 2011, AIED.

[29]  Arthur C. Graesser,et al.  Better to be frustrated than bored: The incidence, persistence, and impact of learners' cognitive-affective states during interactions with three different computer-based learning environments , 2010, Int. J. Hum. Comput. Stud..

[30]  Tzyy-Ping Jung,et al.  EEG Dynamics Reflect the Distinct Cognitive Process of Optic Problem Solving , 2012, PloS one.

[31]  Maria Montessori,et al.  Spontaneous Activity In Education , 1965 .

[32]  Richard J. Murnane,et al.  The New Division of Labor: How Computers Are Creating the Next Job Market , 2004 .

[33]  Kenneth D. Forbus,et al.  CogSketch: Sketch Understanding for Cognitive Science Research and for Education , 2011, Top. Cogn. Sci..

[34]  L. Darling-Hammond,et al.  Prospects and challenges for inquiry-based approaches to learning , 2010 .

[35]  Arthur C. Graesser,et al.  Emote aloud during learning with AutoTutor: Applying the Facial Action Coding System to cognitive–affective states during learning , 2008 .

[36]  Yale Song,et al.  Multimodal human behavior analysis: learning correlation and interaction across modalities , 2012, ICMI '12.

[37]  Clive L. Dym,et al.  Learning Engineering: Design, Languages, and Experiences * , 1999 .

[38]  Bruce L Sherin,et al.  A Computational Study of Commonsense Science: An Exploration in the Automated Analysis of Clinical Interview Data , 2013 .

[39]  Arthur C. Graesser,et al.  Automatic detection of learner’s affect from conversational cues , 2008, User Modeling and User-Adapted Interaction.

[40]  Marcelo Worsley,et al.  Analysing Engineering Expertise of High School Students Using Eye Tracking and Multimodal Learning Analytics , 2013, EDM.

[41]  Kasia Muldner,et al.  "Yes!": Using Tutor and Sensor Data to Predict Moments of Delight during Instructional Activities , 2010, UMAP.

[42]  Daniel C. Richardson,et al.  Looking To Understand: The Coupling Between Speakers' and Listeners' Eye Movements and Its Relationship to Discourse Comprehension , 2005, Cogn. Sci..

[43]  Richard E. Clark,et al.  Why Minimal Guidance During Instruction Does Not Work: An Analysis of the Failure of Constructivist, Discovery, Problem-Based, Experiential, and Inquiry-Based Teaching , 2006 .

[44]  Marcelo Worsley,et al.  What's an Expert? Using Learning Analytics to Identify Emergent Markers of Expertise through Automated Speech, Sentiment and Sketch Analysis , 2011, EDM.

[45]  Diane J. Litman,et al.  The relative impact of student affect on performance models in a spoken dialogue tutoring system , 2008, User Modeling and User-Adapted Interaction.

[46]  Arlinda Cabral,et al.  Pedagogia do Oprimido , 2005 .

[47]  Marcelo Worsley,et al.  Multimodal Learning Analytics as a Tool for Bridging Learning Theory and Complex Learning Behaviors , 2014, MLA@ICMI.

[48]  Dae-Woon Lim,et al.  Changes in Cognitive Performance Due to Three Types of Emotional Tension , 2010, FGIT-DTA/BSBT.

[49]  Gerald Penn,et al.  Improving Automatic Speech Recognition for Lectures through Transformation-based Rules Learned from Minimal Data , 2009, ACL/IJCNLP.

[50]  Cristina Conati,et al.  Empirically building and evaluating a probabilistic model of user affect , 2009, User Modeling and User-Adapted Interaction.

[51]  Niels Henze,et al.  Gesture recognition with a Wii controller , 2008, TEI.

[52]  Bertrand Schneider,et al.  Toward collaboration sensing , 2014, Int. J. Comput. Support. Collab. Learn..

[53]  Maria Caterina Tornatora,et al.  An Eye-Tracking Study of Learning From Science Text With Concrete and Abstract Illustrations , 2013 .

[54]  Ben Cowley,et al.  Cardiovascular physiology predicts learning effects in a serious game activity , 2013, Comput. Educ..

[55]  Rémi Ronfard,et al.  Free viewpoint action recognition using motion history volumes , 2006, Comput. Vis. Image Underst..

[56]  Bertrand Schneider,et al.  3D Tangibles Facilitate Joint Visual Attention in Dyads , 2015, CSCL.

[57]  Joseph E. Beck,et al.  Using Knowledge Tracing in a Noisy Environment to Measure Student Reading Proficiencies , 2006, Int. J. Artif. Intell. Educ..

[58]  J. Dewey The School and Society , 2019 .

[59]  Jack Mostow,et al.  Toward Exploiting EEG Input in a Reading Tutor , 2011, Int. J. Artif. Intell. Educ..

[60]  F. Paas,et al.  Attention guidance in learning from a complex animation: Seeing is understanding? , 2010 .