Reading Students' Multiple Mental States in Conversation from Facial and Heart Rate Cues

Students’ mental states have been widely acknowledged as crucial components for inferring their learning processes and are closely linked with learning outcomes. Understanding students’ complex mental states including concentration, confusion, frustration, and boredom in teacher-student conversation could benefit a human teacher’s perceptual and real-time decision-making capability in providing personalized and adaptive support in coaching activities. Many lines of research have explored the automatic measurement of students’ mental states in pre-designed human-computer tasks. It still remains a challenge to detect the complex mental states of students in real teacher-student conversation. In this study, we made such an attempt by describing a system for predicting the complex mental states of students from multiple perspectives: facial and physiological (heart rate) cues in real student-teacher conversation scenarios. We developed an advanced multi-sensor-based system and applied it in small-scale meetings to collect students’ multimodal conversation data. We demonstrate a multimodal analysis framework. Machine learning models were built by using extracted interpretable proxy features at a fine-grained level to validate their predictive ability regarding students’ multiple mental states. Our results provide evidence of the potential value of fusing multimodal data to understand students’ multiple mental states in real-world student-teacher conversation.

[1]  Eduard Hovy,et al.  Emotion Recognition in Conversation: Research Challenges, Datasets, and Recent Advances , 2019, IEEE Access.

[2]  Katashi Nagao,et al.  Automatic Evaluation of Students' Discussion Skill Based on their Heart Rate , 2018, CSEDU.

[3]  Xiaoyu Shen,et al.  DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset , 2017, IJCNLP.

[4]  Supreeth M. Gowda,et al.  Affective States and State Tests: Investigating How Affect and Engagement during the School Year Predict End-of-Year Learning Outcomes , 2014, J. Learn. Anal..

[5]  Caitlin Mills,et al.  Emotions while writing about emotional and non-emotional topics , 2014 .

[6]  S. D’Mello A selective meta-analysis on the relative incidence of discrete affective states during learning with technology , 2013 .

[7]  Kristy Elizabeth Boyer,et al.  Embodied Affect in Tutorial Dialogue: Student Gesture and Posture , 2013, AIED.

[8]  Kristy Elizabeth Boyer,et al.  Automatically Recognizing Facial Expression: Predicting Engagement and Frustration , 2013, EDM.

[9]  Jelena Obradović,et al.  The construct of psychophysiological reactivity: Statistical and psychometric issues , 2013 .

[10]  J. Bhattacharya,et al.  High-Learners Present Larger Mid-Frontal Theta Power and Connectivity in Response to Incorrect Performance Feedback , 2013, The Journal of Neuroscience.

[11]  Ben Cowley,et al.  Cardiovascular physiology predicts learning effects in a serious game activity , 2013, Comput. Educ..

[12]  Marcelo Worsley,et al.  Analysing Engineering Expertise of High School Students Using Eye Tracking and Multimodal Learning Analytics , 2013, EDM.

[13]  A. Graesser,et al.  Dynamics of affective states during complex learning , 2012 .

[14]  Ryan Shaun Joazeiro de Baker,et al.  The Effects of an Interactive Software Agent on Student Affective Dynamics while Using ;an Intelligent Tutoring System , 2012, IEEE Transactions on Affective Computing.

[15]  Omar AlZoubi,et al.  Affect Detection from Multichannel Physiology during Learning Sessions with AutoTutor , 2011, AIED.

[16]  Marian Stewart Bartlett,et al.  Towards an Optimal Affect-Sensitive Instructional System of cognitive skills , 2011, CVPR 2011 WORKSHOPS.

[17]  Diane J. Litman,et al.  When Does Disengagement Correlate with Learning in Spoken Dialog Computer Tutoring? , 2011, AIED.

[18]  F. Paas,et al.  Attention guidance in learning from a complex animation: Seeing is understanding? , 2010 .

[19]  Rafael A. Calvo,et al.  Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications , 2010, IEEE Transactions on Affective Computing.

[20]  Elaine Toms,et al.  The development and evaluation of a survey to measure user engagement , 2010, J. Assoc. Inf. Sci. Technol..

[21]  Carlos Busso,et al.  IEMOCAP: interactive emotional dyadic motion capture database , 2008, Lang. Resour. Evaluation.

[22]  Arthur C. Graesser,et al.  Automatic detection of learner’s affect from conversational cues , 2008, User Modeling and User-Adapted Interaction.

[23]  Ronald H. Stevens,et al.  EEG-Related Changes in Cognitive Workload, Engagement and Distraction as Students Acquire Problem Solving Skills , 2007, User Modeling.

[24]  Laurence Devillers,et al.  Real-Life Emotion Recognition in Speech , 2007, Speaker Classification.

[25]  Arthur C. Graesser,et al.  Predicting Affective States expressed through an Emote-Aloud Procedure from AutoTutor's Mixed-Initiative Dialogue , 2006, Int. J. Artif. Intell. Educ..

[26]  Ashish Kapoor,et al.  Multimodal affect recognition in learning environments , 2005, ACM Multimedia.

[27]  Lorrie A. Shepard,et al.  Linking Formative Assessment to Scaffolding , 2005 .

[28]  Scotty D. Craig,et al.  Affect and learning: An exploratory look into the role of affect in learning with AutoTutor , 2004 .

[29]  Bronwen Cowie,et al.  The Characteristics of Formative Assessment in Science Education. , 2001 .

[30]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.