MMOD-COG: A Database for Multimodal Cognitive Load Classification

This paper presents a dataset for multimodal classification of cognitive load recorded on a sample of students. The cognitive load was induced by way of performing basic arithmetic tasks, while the multimodal aspect of the dataset comes in the form of both speech and physiological responses to those tasks. The goal of the dataset was two-fold: firstly to provide an alternative to existing cognitive load focused datasets, usually based around Stroop tasks or working memory tasks; and secondly to implement the cognitive load tasks in a way that would make the responses appropriate for both speech and physiological response analysis, ultimately making it multimodal. The paper also presents preliminary classification benchmarks, in which SVM classifiers were trained and evaluated solely on either speech or physiological signals and on combinations of the two. The multimodal nature of the classifiers may provide improvements on results on this inherently challenging machine learning problem because it provides more data about both the intra-participant and inter-participant differences in how cognitive load manifests itself in affective responses.

[1]  Eliathamby Ambikairajah,et al.  Voice source features for cognitive load classification , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[2]  Róbert Busa-Fekete,et al.  Detecting the intensity of cognitive and physical load using AdaBoost and deep rectifier neural networks , 2014, INTERSPEECH.

[3]  Marko Šarlija,et al.  Classification of cognitive load using voice features: A preliminary investigation , 2017, 2017 8th IEEE International Conference on Cognitive Infocommunications (CogInfoCom).

[4]  Carlos Busso,et al.  Iterative Feature Normalization Scheme for Automatic Emotion Detection from Speech , 2013, IEEE Transactions on Affective Computing.

[5]  D. Brodie,et al.  Effects of Short-Term Psychological Stress on the Time and Frequency Domains of Heart-Rate Variability , 2000, Perceptual and motor skills.

[6]  C. Stepp,et al.  Acoustic Measures of Voice and Physiologic Measures of Autonomic Arousal during Speech as a Function of Cognitive Load. , 2017, Journal of voice : official journal of the Voice Foundation.

[7]  Fang Chen,et al.  Automatic cognitive load detection from speech features , 2007, OZCHI '07.

[8]  Vidhyasaharan Sethu,et al.  Investigation of spectral centroid features for cognitive load classification , 2011, Speech Commun..

[9]  Björn W. Schuller,et al.  Recent developments in openSMILE, the munich open-source multimedia feature extractor , 2013, ACM Multimedia.

[10]  Peter Berggren,et al.  Dynamic Assessment of Pilot Mental Status , 2002 .

[11]  Fabien Ringeval,et al.  The INTERSPEECH 2014 computational paralinguistics challenge: cognitive & physical load , 2014, INTERSPEECH.

[12]  Fang Chen,et al.  Speech-based cognitive load monitoring system , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.

[13]  F. Paas,et al.  Cognitive Load Measurement as a Means to Advance Cognitive Load Theory , 2003 .

[14]  Pierre-Vincent Paubel,et al.  Human Voice as a Measure of Mental Load Level. , 2018, Journal of speech, language, and hearing research : JSLHR.

[15]  Eliathamby Ambikairajah,et al.  Formant Frequencies under Cognitive Load: Effects and Classification , 2011, EURASIP J. Adv. Signal Process..

[16]  Jon Gudnason,et al.  Monitoring Cognitive Workload Using Vocal Tract and Voice Source Features , 2017, Period. Polytech. Electr. Eng. Comput. Sci..

[17]  Marko Sarlija,et al.  A convolutional neural network based approach to QRS detection , 2017, Proceedings of the 10th International Symposium on Image and Signal Processing and Analysis.

[18]  Luca Citi,et al.  cvxEDA: A Convex Optimization Approach to Electrodermal Activity Processing , 2016, IEEE Transactions on Biomedical Engineering.

[19]  Shrikanth S. Narayanan,et al.  Classification of cognitive load from speech using an i-vector framework , 2014, INTERSPEECH.

[20]  Björn W. Schuller,et al.  The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing , 2016, IEEE Transactions on Affective Computing.

[21]  F. Shaffer,et al.  An Overview of Heart Rate Variability Metrics and Norms , 2017, Front. Public Health.

[22]  Klaus R. Scherer,et al.  Acoustic correlates of task load and stress , 2002, INTERSPEECH.