End-to-end Facial and Physiological Model for Affective Computing and Applications

In recent years, affective computing and its applications have become a fast-growing research topic. Furthermore, the rise of deep learning has introduced significant improvements in the emotion recognition system compared to classical methods. In this work, we propose a multi-modal emotion recognition model based on deep learning techniques using the combination of peripheral physiological signals and facial expressions. Moreover, we present an improvement to proposed models by introducing latent features extracted from our internal Bio Auto-Encoder (BAE). Both models are trained and evaluated on AMIGOS datasets reporting valence, arousal, and emotion state classification. Finally, to demonstrate a possible medical application in affective computing using deep learning techniques, we applied the proposed method to the assessment of anxiety therapy. To this purpose, a reduced multimodal database has been collected by recording facial expressions and peripheral signals such as electrocardiogram (ECG) and galvanic skin response (GSR) of each patient. Valence and arousal estimates were extracted using our proposed model across the duration of the therapy, with successful evaluation to the different emotional changes in the temporal domain.

[1]  Yu Zhang,et al.  A Survey on Multi-Task Learning , 2017, IEEE Transactions on Knowledge and Data Engineering.

[2]  Terrence J. Sejnowski,et al.  Utilizing Deep Learning Towards Multi-Modal Bio-Sensing and Vision-Based Affective Computing , 2019, IEEE Transactions on Affective Computing.

[3]  K. H. Kim,et al.  Emotion recognition system using short-term monitoring of physiological signals , 2004, Medical and Biological Engineering and Computing.

[4]  Sun Duo,et al.  An E-learning System based on Affective Computing , 2012 .

[5]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Takeo Kanade,et al.  Recognizing Action Units for Facial Expression Analysis , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[7]  Mohammed Yeasin,et al.  Recognition of facial expressions and measurement of levels of interest from video , 2006, IEEE Transactions on Multimedia.

[8]  Jiwen Lu,et al.  Single Sample Face Recognition via Learning Deep Supervised Autoencoders , 2015, IEEE Transactions on Information Forensics and Security.

[9]  Zhong Yin,et al.  Recognition of emotions using multimodal physiological signals and an ensemble deep learning model , 2017, Comput. Methods Programs Biomed..

[10]  Lan Li,et al.  Emotion Recognition Using Physiological Signals from Multiple Subjects , 2006, 2006 International Conference on Intelligent Information Hiding and Multimedia.

[11]  Xavier Binefa,et al.  Fully End-to-End Composite Recurrent Convolution Network for Deformable Facial Tracking In The Wild , 2019, 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019).

[12]  Lucas Theis,et al.  Lossy Image Compression with Compressive Autoencoders , 2017, ICLR.

[13]  Chi-Chun Lee,et al.  An Attribute-invariant Variational Learning for Emotion Recognition Using Physiology , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[14]  Christine L. Lisetti,et al.  Using Noninvasive Wearable Computers to Recognize Human Emotions from Physiological Signals , 2004, EURASIP J. Adv. Signal Process..

[15]  James L.Oschman,et al.  Energy Medicine: The Scientific Basis , 2000 .

[16]  P. Lang The emotion probe. Studies of motivation and attention. , 1995, The American psychologist.

[17]  Jennifer Healey,et al.  Toward Machine Emotional Intelligence: Analysis of Affective Physiological State , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[18]  Antoinette M. Feleky The expression of the emotions. , 1914 .

[19]  P. Ekman,et al.  Autonomic nervous system activity distinguishes among emotions. , 1983, Science.

[20]  W. Cannon The James-Lange theory of emotions: a critical examination and an alternative theory. By Walter B. Cannon, 1927. , 1927, The American journal of psychology.

[21]  Samit Bhattacharya,et al.  Using Deep and Convolutional Neural Networks for Accurate Emotion Classification on DEAP Dataset , 2017, AAAI.

[22]  J. Russell A circumplex model of affect. , 1980 .

[23]  Luca Antiga,et al.  Automatic differentiation in PyTorch , 2017 .

[24]  Enas Abdulhay,et al.  Using Deep Convolutional Neural Network for Emotion Detection on a Physiological Signals Dataset (AMIGOS) , 2019, IEEE Access.

[25]  Jason Williams,et al.  Emotion Recognition Using Bio-sensors: First Steps towards an Automatic System , 2004, ADS.

[26]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[28]  U. Rajendra Acharya,et al.  An efficient compression of ECG signals using deep convolutional autoencoders , 2018, Cognitive Systems Research.

[29]  Dong Keun Kim,et al.  Arousal and Valence Classification Model Based on Long Short-Term Memory and DEAP Data for Mental Healthcare Management , 2018, Healthcare informatics research.

[30]  Changchun Liu,et al.  Online Affect Detection and Robot Behavior Adaptation for Intervention of Children With Autism , 2008, IEEE Transactions on Robotics.

[31]  M. Cabanac What is emotion? , 2002, Behavioural Processes.

[32]  Thierry Pun,et al.  DEAP: A Database for Emotion Analysis ;Using Physiological Signals , 2012, IEEE Transactions on Affective Computing.

[33]  Thierry Pun,et al.  Multimodal Emotion Recognition in Response to Videos , 2012, IEEE Transactions on Affective Computing.

[34]  W. Cannon The James-Lange theory of emotions: a critical examination and an alternative theory. By Walter B. Cannon, 1927. , 1927, American Journal of Psychology.

[35]  Mohammad H. Mahoor,et al.  A wavelet-based approach to emotion classification using EDA signals , 2018, Expert Syst. Appl..

[36]  Wei Liu,et al.  Multimodal Emotion Recognition Using Deep Neural Networks , 2017, ICONIP.

[37]  Michael A. Casey,et al.  Musical Audio Synthesis Using Autoencoding Neural Nets , 2014, ICMC.

[38]  M. Kramer Nonlinear principal component analysis using autoassociative neural networks , 1991 .

[39]  Zhang Xiong,et al.  3D object retrieval with stacked local convolutional autoencoder , 2015, Signal Process..

[40]  Salim Lahmiri,et al.  A weighted bio-signal denoising approach using empirical mode decomposition , 2015, Biomedical Engineering Letters.

[41]  Nicu Sebe,et al.  AMIGOS: A dataset for Mood, personality and affect research on Individuals and GrOupS , 2017, ArXiv.

[42]  Guoying Zhao,et al.  Deep Affect Prediction in-the-Wild: Aff-Wild Database and Challenge, Deep Architectures, and Beyond , 2018, International Journal of Computer Vision.

[43]  Honglak Lee,et al.  Deep learning for robust feature generation in audiovisual emotion recognition , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[44]  E. Vesterinen,et al.  Affective Computing , 2009, Encyclopedia of Biometrics.

[45]  Luca Chittaro,et al.  Affective computing vs. affective placebo: Study of a biofeedback-controlled game for relaxation training , 2014, Int. J. Hum. Comput. Stud..

[46]  ZhangJianhua,et al.  Recognition of emotions using multimodal physiological signals and an ensemble deep learning model , 2017 .

[47]  Matjaz Gams,et al.  An Inter-domain Study for Arousal Recognition from Physiological Signals , 2018, Informatica.

[48]  C. Darwin The Expression of the Emotions in Man and Animals , .

[49]  Gyanendra K. Verma,et al.  Multimodal fusion framework: A multiresolution approach for emotion classification and recognition from physiological signals , 2014, NeuroImage.

[50]  King-Sun Fu,et al.  IEEE Transactions on Pattern Analysis and Machine Intelligence Publication Information , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.