CorrNet: Fine-Grained Emotion Recognition for Video Watching Using Wearable Physiological Sensors

Recognizing user emotions while they watch short-form videos anytime and anywhere is essential for facilitating video content customization and personalization. However, most works either classify a single emotion per video stimuli, or are restricted to static, desktop environments. To address this, we propose a correlation-based emotion recognition algorithm (CorrNet) to recognize the valence and arousal (V-A) of each instance (fine-grained segment of signals) using only wearable, physiological signals (e.g., electrodermal activity, heart rate). CorrNet takes advantage of features both inside each instance (intra-modality features) and between different instances for the same video stimuli (correlation-based features). We first test our approach on an indoor-desktop affect dataset (CASE), and thereafter on an outdoor-mobile affect dataset (MERCA) which we collected using a smart wristband and wearable eyetracker. Results show that for subject-independent binary classification (high-low), CorrNet yields promising recognition accuracies: 76.37% and 74.03% for V-A on CASE, and 70.29% and 68.15% for V-A on MERCA. Our findings show: (1) instance segment lengths between 1–4 s result in highest recognition accuracies (2) accuracies between laboratory-grade and wearable sensors are comparable, even under low sampling rates (≤64 Hz) (3) large amounts of neutral V-A labels, an artifact of continuous affect annotation, result in varied recognition performance.

[1]  Sten Hanke,et al.  Emotion Recognition from Physiological Signal Analysis: A Review , 2019, BRAINS/WS-AFFIN@AmI.

[2]  Kriti Sethi,et al.  Stress detection and relief using wearable physiological sensors , 2019 .

[3]  Jeffrey M Girard,et al.  DARMA: Software for dual axis rating and media annotation , 2017, Behavior Research Methods.

[4]  Mohammad Soleymani,et al.  Analysis of EEG Signals and Facial Expressions for Continuous Emotion Detection , 2016, IEEE Transactions on Affective Computing.

[5]  M. Bradley,et al.  Measuring emotion: the Self-Assessment Manikin and the Semantic Differential. , 1994, Journal of behavior therapy and experimental psychiatry.

[6]  Bernd Resch,et al.  Detecting Moments of Stress from Measurements of Wearable Physiological Sensors , 2019, Sensors.

[7]  Alice H. Oh,et al.  K-EmoCon, a multimodal sensor dataset for continuous emotion recognition in naturalistic conversations , 2020, Scientific Data.

[8]  Mohammad Soleymani,et al.  Continuous emotion detection using EEG signals and facial expressions , 2014, 2014 IEEE International Conference on Multimedia and Expo (ICME).

[9]  Alexander J Casson,et al.  Wearable EEG and beyond , 2019, Biomedical Engineering Letters.

[10]  Richard W. Daniels,et al.  Approximation Methods for Electronic Filter Design: With Applications to Passive, Active and Digital Networks , 1974 .

[11]  Arun Ross,et al.  Score normalization in multimodal biometric systems , 2005, Pattern Recognit..

[12]  R. Levenson Emotion and the autonomic nervous system: A prospectus for research on autonomic specificity. , 1988 .

[13]  Thierry Pun,et al.  Multimodal Emotion Recognition in Response to Videos , 2012, IEEE Transactions on Affective Computing.

[14]  Aboul Ella Hassanien,et al.  Intelligent human emotion recognition based on elephant herding optimization tuned support vector regression , 2018, Biomed. Signal Process. Control..

[15]  Jeff A. Bilmes,et al.  Deep Canonical Correlation Analysis , 2013, ICML.

[16]  Hirokazu Tanaka,et al.  Blood pressure estimation from pulse wave velocity measured on the chest , 2013, 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[17]  Elena Di Lascio,et al.  Using Unobtrusive Wearable Sensors to Measure the Physiological Synchrony Between Presenters and Audience Members , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[18]  John A. Stankovic,et al.  A Weakly Supervised Learning Framework for Detecting Social Anxiety and Depression , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[19]  Zhu Wang,et al.  EmotionSense: Emotion Recognition Based on Wearable Wristband , 2018, 2018 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI).

[20]  Elena Di Lascio,et al.  Unobtrusive Assessment of Students' Emotional Engagement during Lectures Using Electrodermal Activity Sensors , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[21]  Gary E. Birch,et al.  Comparison of Evaluation Metrics in Classification Applications with Imbalanced Datasets , 2008, 2008 Seventh International Conference on Machine Learning and Applications.

[22]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[23]  Wioleta Szwoch,et al.  Using physiological signals for emotion recognition , 2013, International Conference on Human System Interaction.

[24]  Sazali Yaacob,et al.  Descriptive Analysis of Skin Temperature Variability of Sympathetic Nervous System Activity in Stress , 2012 .

[25]  Andrea Pinna,et al.  Towards Better Understanding of Player's Game Experience , 2018, ICMR.

[26]  P. Ekman Emotions Revealed: Recognizing Faces and Feelings to Improve Communication and Emotional Life , 2003 .

[27]  Chuan-Yu Chang,et al.  Based on Support Vector Regression for emotion recognition using physiological signals , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[28]  Trisha T. C. Lin Investigating adopter categories and determinants affecting the adoption of mobile television in China 1 , 2014 .

[29]  Sonia H. Contreras Ortiz,et al.  A machine learning model for emotion recognition from physiological signals , 2020, Biomed. Signal Process. Control..

[30]  Rosalind W. Picard Future affective technology for autism and emotion communication , 2009, Philosophical Transactions of the Royal Society B: Biological Sciences.

[31]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[32]  Zhenqi Li,et al.  A Review of Emotion Recognition Using Physiological Signals , 2018, Sensors.

[33]  Adel Oulefki,et al.  Low-Light Face Image Enhancement Based on Dynamic Face Part Selection , 2019, IbPRIA.

[34]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.

[35]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[36]  H. Critchley,et al.  Interoception and emotion. , 2017, Current opinion in psychology.

[37]  J. van Leeuwen,et al.  Neural Networks: Tricks of the Trade , 2002, Lecture Notes in Computer Science.

[38]  Christos D. Katsis,et al.  A User Independent, Biosignal Based, Emotion Recognition Method , 2007, User Modeling.

[39]  Elena Di Lascio,et al.  Detection of Artifacts in Ambulatory Electrodermal Activity Data , 2020, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[40]  Luca Romeo,et al.  Multiple Instance Learning for Emotion Recognition Using Physiological Signals , 2022, IEEE Transactions on Affective Computing.

[41]  Eckart Altenmüller,et al.  EMuJoy: Software for continuous measurement of perceived emotions in music , 2007, Behavior research methods.

[42]  Jürgen Schmidhuber,et al.  LSTM: A Search Space Odyssey , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[43]  Minghao Wang,et al.  Multi-Feature Based Emotion Recognition for Video Clips , 2018, ICMI.

[44]  Lutz Prechelt,et al.  Early Stopping - But When? , 2012, Neural Networks: Tricks of the Trade.

[45]  Nico Pallamin,et al.  Emotion Recognition Using Physiological Signals: Laboratory vs. Wearable Sensors , 2017, AHFE.

[46]  Niels Henze,et al.  Understanding Face and Eye Visibility in Front-Facing Cameras of Smartphones used in the Wild , 2018, CHI.

[47]  Yunhong Wang,et al.  Continuous Emotion Recognition in Videos by Fusing Facial Expression, Head Pose and Eye Gaze , 2019, ICMI.

[48]  Tianyi Zhang,et al.  Multi-modal Fusion Methods for Robust Emotion Recognition using Body-worn Physiological Sensors in Mobile Environments , 2019, ICMI.

[49]  P. Dario,et al.  Control of multifunctional prosthetic hands by processing the electromyographic signal. , 2002, Critical reviews in biomedical engineering.

[50]  Markus Löchtefeld,et al.  Face mask design to mitigate facial expression occlusion , 2020, SEMWEB.

[51]  Michael R. Ibbotson,et al.  Saccadic Modulation of Neural Responses: Possible Roles in Saccadic Suppression, Enhancement, and Time Compression , 2008, The Journal of Neuroscience.

[52]  Anil A. Bharath,et al.  On denoising autoencoders trained to minimise binary cross-entropy , 2017, ArXiv.

[53]  Lei Zhou,et al.  Using Physiological Measures to Evaluate User Experience of Mobile Applications , 2014, HCI.

[54]  Tomás Lozano-Pérez,et al.  A Framework for Multiple-Instance Learning , 1997, NIPS.

[55]  E. Meijering A chronology of interpolation: from ancient astronomy to modern signal and image processing , 2002, Proc. IEEE.

[56]  Sheng-hua Zhong,et al.  Implicit Affective Video Tagging Using Pupillary Response , 2018, MMM.

[57]  Touradj Ebrahimi,et al.  Affect recognition based on physiological changes during the watching of music videos , 2012, TIIS.

[58]  Tapio Seppänen,et al.  Enhancing Emotion Recognition from ECG Signals using Supervised Dimensionality Reduction , 2017, ICPRAM.

[59]  Ram Mohana Reddy Guddeti,et al.  EmoWare: A Context-Aware Framework for Personalized Video Recommendation Using Affective Video Sequences , 2019, IEEE Access.

[60]  Batya Friedman,et al.  The Watcher and the Watched: Social Judgments About Privacy in a Public Place , 2006, Media Space 20+ Years of Mediated Life.

[61]  Hao Tang,et al.  Emotion Recognition using Multimodal Residual LSTM Network , 2019, ACM Multimedia.

[62]  Pablo César,et al.  CorrFeat: Correlation-based Feature Extraction Algorithm using Skin Conductance and Pupil Diameter for Emotion Recognition , 2019, ICMI.

[63]  Thierry Pun,et al.  DEAP: A Database for Emotion Analysis ;Using Physiological Signals , 2012, IEEE Transactions on Affective Computing.

[64]  P. Ekman An argument for basic emotions , 1992 .

[65]  Jianda Han,et al.  Physiological Signal-Based Method for Measurement of Pain Intensity , 2017, Front. Neurosci..

[66]  Sahar Moghimi,et al.  Continuous Emotion Recognition during Music Listening Using EEG Signals: A Fuzzy Parallel Cascades Model , 2019, Appl. Soft Comput..

[67]  Friedhelm Schwenker,et al.  A dataset of continuous affect annotations and physiological signals for emotion analysis , 2018, Scientific Data.

[68]  Witold Pedrycz,et al.  Adaptive Feature Selection-Based AdaBoost-KNN With Direct Optimization for Dynamic Emotion Recognition in Human–Robot Interaction , 2019, IEEE Transactions on Emerging Topics in Computational Intelligence.

[69]  C. L. Philip Chen,et al.  Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[70]  Daniel P. Siewiorek,et al.  Using physiological sensors to detect levels of user frustration induced by system delays , 2015, UbiComp.

[71]  M Murugappan,et al.  Physiological signals based human emotion Recognition: a review , 2011, 2011 IEEE 7th International Colloquium on Signal Processing and its Applications.

[72]  Yongtian He,et al.  Deep learning for electroencephalogram (EEG) classification tasks: a review , 2019, Journal of neural engineering.

[73]  Wei Xu,et al.  Bidirectional LSTM-CRF Models for Sequence Tagging , 2015, ArXiv.

[74]  Athanasios Katsamanis,et al.  Affective State Recognition in Married Couples' Interactions Using PCA-Based Vocal Entrainment Measures with Multiple Instance Learning , 2011, ACII.

[75]  Marko Horvat,et al.  Comparative analysis of emotion estimation methods based on physiological measurements for real-time applications , 2014, Int. J. Hum. Comput. Stud..

[76]  Hugo Larochelle,et al.  An Autoencoder Approach to Learning Bilingual Word Representations , 2014, NIPS.

[77]  Kristof Van Laerhoven,et al.  Labelling Affective States "in the Wild": Practical Guidelines and Lessons Learned , 2018, UbiComp/ISWC Adjunct.

[78]  Carlos Carrascosa,et al.  Emotions Detection on an Ambient Intelligent System Using Wearable Devices , 2019, AfCAI.

[79]  Najlae Idrissi,et al.  Driver Fatigue Detection using Recurrent Neural Networks , 2019, NISS19.

[80]  Qiang Yang,et al.  Music Emotion Recognition by Multi-label Multi-layer Multi-instance Multi-view Learning , 2014, ACM Multimedia.

[81]  Zhen-Hua Ling,et al.  Enhanced LSTM for Natural Language Inference , 2016, ACL.

[82]  Hui Yu,et al.  Realistic Facial Expression Reconstruction for VR HMD Users , 2020, IEEE Transactions on Multimedia.

[83]  Jianmin Jiang,et al.  An Attentional-LSTM for Improved Classification of Brain Activities Evoked by Images , 2019, ACM Multimedia.

[84]  Rose T. Faghih,et al.  Online and offline anger detection via electromyography analysis , 2017, 2017 IEEE Healthcare Innovations and Point of Care Technologies (HI-POCT).

[85]  Kyandoghere Kyamakya,et al.  A Globally Generalized Emotion Recognition System Involving Different Physiological Signals , 2018, Sensors.

[86]  James L. Coyle,et al.  Deep Belief Networks for Electroencephalography: A Review of Recent Contributions and Future Outlooks , 2017, IEEE Journal of Biomedical and Health Informatics.

[87]  Steven X. Ding,et al.  A Broad Learning Aided Data-Driven Framework of Fast Fault Diagnosis for High-Speed Trains , 2021, IEEE Intelligent Transportation Systems Magazine.

[88]  Roddy Cowie,et al.  FEELTRACE: an instrument for recording perceived emotion in real time , 2000 .

[89]  Mohammad Soleymani,et al.  A Multimodal Database for Affect Recognition and Implicit Tagging , 2012, IEEE Transactions on Affective Computing.

[90]  Taryn L. Stanko,et al.  Watching You Watching Me: Boundary Control and Capturing Attention in the Context of Ubiquitous Technology Use , 2014 .

[91]  Shrikanth Narayanan,et al.  Multiple Instance Learning for Behavioral Coding , 2017, IEEE Transactions on Affective Computing.

[92]  Johannes Hewig,et al.  A revised film set for the induction of basic emotions. , 2005 .

[93]  Cheng He,et al.  An Emotion Recognition System Based on Physiological Signals Obtained by Wearable Sensors , 2017 .

[94]  Subramanian Ramanathan,et al.  DECAF: MEG-Based Multimodal Database for Decoding Affective Physiological Responses , 2015, IEEE Transactions on Affective Computing.

[95]  Matthew D. Zeiler ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.

[96]  W. Larry Kenney,et al.  Sympathetic regulation during thermal stress in human aging and disease , 2016, Autonomic Neuroscience.

[97]  Frank Bentley,et al.  Understanding Mass-Market Mobile TV Behaviors in the Streaming Era , 2019, CHI.

[98]  Enas Abdulhay,et al.  Using Deep Convolutional Neural Network for Emotion Detection on a Physiological Signals Dataset (AMIGOS) , 2019, IEEE Access.

[99]  Fadi Al Machot,et al.  CNN Based Subject-Independent Driver Emotion Recognition System Involving Physiological Signals for ADAS , 2016 .

[100]  P. Lang The emotion probe. Studies of motivation and attention. , 1995, The American psychologist.

[101]  Hatice Gunes,et al.  Continuous Prediction of Spontaneous Affect from Multiple Cues and Modalities in Valence-Arousal Space , 2011, IEEE Transactions on Affective Computing.

[102]  Nancy Chinchor,et al.  MUC-3 evaluation metrics , 1991, MUC.

[103]  Yoshihiko Suhara,et al.  DeepMood: Forecasting Depressed Mood Based on Self-Reported Histories via Recurrent Neural Networks , 2017, WWW.

[104]  Hongxun Yao,et al.  Predicting Continuous Probability Distribution of Image Emotions in Valence-Arousal Space , 2015, ACM Multimedia.

[105]  Chen Wang,et al.  RCEA: Real-time, Continuous Emotion Annotation for Collecting Precise Mobile Video Ground Truth Labels , 2020, CHI.

[106]  Julien Fleureau,et al.  Affective Benchmarking of Movies Based on the Physiological Responses of a Real Audience , 2013, 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction.

[107]  Kenton O'Hara,et al.  Consuming video on mobile devices , 2007, CHI.

[108]  Ellen Elizabeth Bartolini,et al.  Eliciting Emotion with Film: Development of a Stimulus Set , 2011 .

[109]  Mahesh Sooriyabandara,et al.  HealthyOffice: Mood recognition at work using smartphones and wearable sensors , 2016, 2016 IEEE International Conference on Pervasive Computing and Communication Workshops (PerCom Workshops).

[110]  D. Olson,et al.  Circumplex Model: Systemic Assessment and Treatment of Families , 1989 .

[111]  F Draicchio,et al.  Oculomotor nerve regeneration in rats. Functional, histological, and neuroanatomical studies. , 1987, Journal of neurosurgery.

[112]  Mustafa E. Kamasak,et al.  Emotion Based Music Recommendation System Using Wearable Physiological Sensors , 2018, IEEE Transactions on Consumer Electronics.

[113]  Jurij F. Tasic,et al.  Affective Labeling in a Content-Based Recommender System for Images , 2013, IEEE Transactions on Multimedia.

[114]  Jennifer McNally,et al.  How Millennials and Teens Consume Mobile Video , 2017, TVX.

[115]  Xiaowei Niu,et al.  Emotion Pattern Recognition Using Physiological Signals , 2014 .

[116]  Jie Wei,et al.  Higher-order Multivariable Polynomial Regression to Estimate Human Affective States , 2016, Scientific Reports.

[117]  Mohsen Guizani,et al.  Deep Feature Learning for Medical Image Analysis with Convolutional Autoencoder Neural Network , 2017, IEEE Transactions on Big Data.

[118]  Jianwei Niu,et al.  Affivir: An affect-based Internet video recommendation system , 2013, Neurocomputing.

[119]  J. Russell A circumplex model of affect. , 1980 .

[120]  Rafael A. Calvo,et al.  Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications , 2010, IEEE Transactions on Affective Computing.