Facial Expression and Peripheral Physiology Fusion to Decode Individualized Affective Experience

In this paper, we present a multimodal approach to simultaneously analyze facial movements and several peripheral physiological signals to decode individualized affective experiences under positive and negative emotional contexts, while considering their personalized resting dynamics. We propose a person-specific recurrence network to quantify the dynamics present in the person's facial movements and physiological data. Facial movement is represented using a robust head vs. 3D face landmark localization and tracking approach, and physiological data are processed by extracting known attributes related to the underlying affective experience. The dynamical coupling between different input modalities is then assessed through the extraction of several complex recurrent network metrics. Inference models are then trained using these metrics as features to predict individual's affective experience in a given context, after their resting dynamics are excluded from their response. We validated our approach using a multimodal dataset consists of (i) facial videos and (ii) several peripheral physiological signals, synchronously recorded from 12 participants while watching 4 emotion-eliciting video-based stimuli. The affective experience prediction results signified that our multimodal fusion method improves the prediction accuracy up to 19% when compared to the prediction using only one or a subset of the input modalities. Furthermore, we gained prediction improvement for affective experience by considering the effect of individualized resting dynamics.

[1]  Aleix M. Martínez,et al.  EmotioNet: An Accurate, Real-Time Algorithm for the Automatic Annotation of a Million Facial Expressions in the Wild , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[3]  Mohsen Nabian,et al.  Analysis of Multimodal Physiological Signals Within and Across Individuals to Predict Psychological Threat vs. Challenge , 2017 .

[4]  Hong Jin Joo,et al.  The Impact of Personality Traits on Emotional Responses to Interpersonal Stress , 2012, Clinical psychopharmacology and neuroscience : the official scientific journal of the Korean College of Neuropsychopharmacology.

[5]  Mikhail Kuznetsov,et al.  Filtering the surface EMG signal: Movement artifact and baseline noise contamination. , 2010, Journal of biomechanics.

[6]  F. Takens Detecting strange attractors in turbulence , 1981 .

[7]  Maja Pantic,et al.  The first facial expression recognition and analysis challenge , 2011, Face and Gesture 2011.

[8]  Catherine Nielson Extracting Facial Synchrony from Videos of Naturalistic Dyadic Interaction , 2018 .

[9]  Benjamin Pfaff,et al.  Theories Of Human Communication , 2016 .

[10]  P. Ekman Facial expression and emotion. , 1993, The American psychologist.

[11]  Jing Xiao,et al.  Automatic analysis and recognition of brow actions and head motion in spontaneous facial behavior , 2004, 2004 IEEE International Conference on Systems, Man and Cybernetics (IEEE Cat. No.04CH37583).

[12]  J. Fernández-Dols,et al.  Emotion and Expression: Naturalistic Studies , 2013 .

[13]  L. F. Barrett,et al.  Affect as a Psychological Primitive. , 2009, Advances in experimental social psychology.

[14]  Takeo Kanade,et al.  Recognizing Action Units for Facial Expression Analysis , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[15]  A. L. N. Fred,et al.  An Electrodermal Activity Psychophysiologic Model , 2007 .

[16]  Rameshwari S Mane,et al.  Cardiac Arrhythmia Detection By ECG Feature Extraction , 2013 .

[17]  Rosalind W. Picard Toward computers that recognize and respond to user emotion , 2000, IBM Syst. J..

[18]  J. Russell,et al.  The science of facial expression. , 2017 .

[19]  L. F. Barrett,et al.  An active inference theory of allostasis and interoception in depression , 2016, Philosophical Transactions of the Royal Society B: Biological Sciences.

[20]  Murat Akçakaya,et al.  Decoding emotional experiences through physiological signal processing , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[21]  D. Ruelle,et al.  Recurrence Plots of Dynamical Systems , 1987 .

[22]  Maja Pantic,et al.  Dynamics of facial expression: recognition of facial actions and their temporal segments from face profile image sequences , 2006, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[23]  Ioannis Patras,et al.  Fusion of facial expressions and EEG for implicit affective tagging , 2013, Image Vis. Comput..

[24]  Marieke Wichers,et al.  From Affective Experience to Motivated Action: Tracking Reward-Seeking and Punishment-Avoidant Behaviour in Real-Life , 2015, PloS one.

[25]  J. Panksepp Affective Neuroscience: The Foundations of Human and Animal Emotions , 1998 .

[26]  Yuxiao Hu,et al.  Spontaneous Emotional Facial Expression Detection , 2006, J. Multim..

[27]  Gwen Littlewort,et al.  Fully Automatic Facial Action Recognition in Spontaneous Behavior , 2006, 7th International Conference on Automatic Face and Gesture Recognition (FGR06).

[28]  Fabien Ringeval,et al.  AV+EC 2015: The First Affect Recognition Challenge Bridging Across Audio, Video, and Physiological Data , 2015, AVEC@ACM Multimedia.

[29]  Mohsen Nabian,et al.  An Open-Source Feature Extraction Tool for the Analysis of Peripheral Physiological Data , 2018, IEEE Journal of Translational Engineering in Health and Medicine.

[30]  Gyanendra K. Verma,et al.  Multimodal fusion framework: A multiresolution approach for emotion classification and recognition from physiological signals , 2014, NeuroImage.

[31]  D. Ruppert The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2004 .

[32]  Alistair A. Young,et al.  Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) , 2017, MICCAI 2017.

[33]  Zhiwei Zhu,et al.  Toward a decision-theoretic framework for affect recognition and user assistance , 2006, Int. J. Hum. Comput. Stud..

[34]  J. Russell Is there universal recognition of emotion from facial expression? A review of the cross-cultural studies. , 1994, Psychological bulletin.

[35]  Guodong Guo,et al.  Learning from examples in the small sample case: face expression recognition , 2005, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[36]  Josephine Sullivan,et al.  One millisecond face alignment with an ensemble of regression trees , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[37]  Zhihong Zeng,et al.  A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[38]  Javier Hernandez,et al.  Call Center Stress Recognition with Person-Specific Models , 2011, ACII.

[39]  Rainer Reisenzein,et al.  Coherence between emotions and facial expressions: a research synthesis , 2017 .

[40]  Vladimir Pavlovic,et al.  Personalized Modeling of Facial Action Unit Intensity , 2014, ISVC.

[41]  Mohsen Nabian,et al.  A biosignal-specific processing tool for machine learning and pattern recognition , 2017, 2017 IEEE Healthcare Innovations and Point of Care Technologies (HI-POCT).

[42]  KandemirMelih,et al.  Multi-task and multi-view learning of user state , 2014 .

[43]  Chun-An Chou,et al.  Recognizing affective state patterns using regularized learning with nonlinear dynamical features of EEG , 2018, 2018 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI).

[44]  Tanja Schultz,et al.  Towards emotion recognition from electroencephalographic signals , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[45]  William J. Christmas,et al.  3D Morphable Face Models and Their Applications , 2016, AMDO.

[46]  Jing Xiao,et al.  Robust full-motion recovery of head by dynamic templates and re-registration techniques , 2002, Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition.

[47]  Lijun Yin,et al.  FERA 2015 - second Facial Expression Recognition and Analysis challenge , 2015, 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[48]  Jürgen Kurths,et al.  Multivariate recurrence plots , 2004 .

[49]  S. Kühn,et al.  What goes on in the resting-state? A qualitative glimpse into resting-state experience in the scanner , 2015, Front. Psychol..

[50]  J. Kurths,et al.  Identifying coupling directions by recurrences , 2015 .

[51]  Fridanna Maricchiolo,et al.  Be Careful Where You Smile: Culture Shapes Judgments of Intelligence and Honesty of Smiling Individuals , 2015, Journal of Nonverbal Behavior.

[52]  Michel F. Valstar,et al.  Cascaded Continuous Regression for Real-time Incremental Face Tracking , 2016, ECCV.

[53]  Mahadev D. Uplane,et al.  Digital elliptic filter application for noise reduction in ECG signal , 2005 .

[54]  Olaf Sporns,et al.  Complex network measures of brain connectivity: Uses and interpretations , 2010, NeuroImage.

[55]  G. Horstmann,et al.  Coherence between Emotion and Facial Expression: Evidence from Laboratory Experiments , 2013 .

[56]  Willis J. Tompkins,et al.  A Real-Time QRS Detection Algorithm , 1985, IEEE Transactions on Biomedical Engineering.

[57]  Samuel Kaski,et al.  Multi-task and multi-view learning of user state , 2014, Neurocomputing.