Emotional facial sensing and multimodal fusion in a continuous 2D affective space

This paper deals with two main research focuses on Affective Computing: facial emotion recognition and multimodal fusion of affective information coming from different channels. The facial sensing system developed implements an emotional classification mechanism that combines, in a novel and robust manner, the five most commonly used classifiers in the field of affect sensing, obtaining at the output an associated weight of the facial expression to each of the six Ekman’s universal emotional categories plus the neutral. The system is able to analyze any subject, male or female, of any age, and ethnicity and has been validated by means of statistical evaluation strategies, such as cross-validation, classification accuracy ratios and confusion matrices. The categorical facial sensing system has been subsequently expanded to a continuous 2D affective space which has made it also possible to face the problem of multimodal human affect recognition. A novel fusion methodology able to fuse any number of affective modules, with very different time-scales and output labels, is proposed. It relies on the 2D Whissell affective space and is able to output a continuous emotional path characterizing the user’s affective progress over time. A Kalman filtering technique controls this path in real-time to ensure temporal consistency and robustness to the system. Moreover, the methodology is adaptive to eventual temporal changes in the reliability of the different inputs’ quality. The potential of the multimodal fusion methodology is demonstrated by fusing dynamic affective information extracted from different channels (video, typed-in text and emoticons) of an Instant Messaging tool.

[1]  Ipke Wachsmuth,et al.  Simulating empathy for the virtual human Max , 2007 .

[2]  Qiang Ji,et al.  A probabilistic framework for modeling and real-time monitoring human fatigue , 2006, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[3]  J. G. Taylor,et al.  Emotion recognition in human-computer interaction , 2005, Neural Networks.

[4]  Marc Cavazza,et al.  PAD-based multimodal affective fusion , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[5]  Frédéric Gosselin,et al.  Bubbles: a technique to reveal the use of information in recognition tasks , 2001, Vision Research.

[6]  Susan R. Fussell,et al.  Cultural differences in the use of instant messaging in Asia and North America , 2006, CSCW '06.

[7]  A. Freitas-Magalhães Facial Expression of Emotion , 2012 .

[8]  Gwen Littlewort,et al.  Dynamics of Facial Expression Extracted Automatically from Video , 2004, CVPR Workshops.

[9]  J. N. Bassili Emotion recognition: the role of facial movement and the relative importance of upper and lower areas of the face. , 1979, Journal of personality and social psychology.

[10]  Zakia Hammal,et al.  Belief Theory Applied to Facial Expressions Classification , 2005, ICAPR.

[11]  FragopanagosN.,et al.  2005 Special Issue , 2005 .

[12]  Zhihong Zeng,et al.  Audio-Visual Affect Recognition , 2007, IEEE Transactions on Multimedia.

[13]  Shaogang Gong,et al.  Beyond Facial Expressions: Learning Human Emotion from Body Gestures , 2007, BMVC.

[14]  J. Alfredo Sánchez,et al.  Conveying mood and emotion in instant messaging by using a two-dimensional model for affective states , 2006, IHC '06.

[15]  Hatice Gunes,et al.  From the Lab to the real world: affect recognition using multiple cues and modalities , 2008 .

[16]  อนิรุธ สืบสิงห์,et al.  Data Mining Practical Machine Learning Tools and Techniques , 2014 .

[17]  Mohammed Yeasin,et al.  Recognition of facial expressions and measurement of levels of interest from video , 2006, IEEE Transactions on Multimedia.

[18]  Maja Pantic,et al.  Web-based database for facial expression analysis , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[19]  S. Voloshynovskiy,et al.  Brain-computer interaction research at the computer vision and multimedia laboratory, University of Geneva , 2006, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[20]  Piyush Kumar,et al.  Minimum-Volume Enclosing Ellipsoids and Core Sets , 2005 .

[21]  Hasan Demirel,et al.  Facial Expression Recognition Using 3D Facial Feature Distances , 2007, ICIAR.

[22]  Alecia Wolf,et al.  Emotional Expression Online: Gender Differences in Emoticon Use , 2000, Cyberpsychology Behav. Soc. Netw..

[23]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[24]  Eva Cerezo,et al.  Effective Emotional Classification Combining Facial Classifiers and User Assessment , 2008, AMDO.

[25]  Darryl Morrell,et al.  An Extended Set-valued Kalman Filter , 2003, ISIPTA.

[26]  E. Vesterinen,et al.  Affective Computing , 2009, Encyclopedia of Biometrics.

[27]  R. Plutchik Emotion, a psychoevolutionary synthesis , 1980 .

[28]  Mark A. Hall,et al.  Correlation-based Feature Selection for Machine Learning , 2003 .

[29]  Zhihong Zeng,et al.  A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions , 2009, IEEE Trans. Pattern Anal. Mach. Intell..

[30]  Rosalind W. Picard Affective computing: (526112012-054) , 1997 .

[31]  Kostas Karpouzis,et al.  The HUMAINE Database: Addressing the Collection and Annotation of Naturalistic and Induced Emotional Data , 2007, ACII.

[32]  T. Dalgleish,et al.  Handbook of Cognition and Emotion: Dalgleish/Cognition and Emotion , 2005 .

[33]  Loïc Kessous,et al.  Modeling naturalistic affective states via facial and vocal expressions recognition , 2006, ICMI '06.

[34]  Renaud Séguier,et al.  Automatic design of a control interface for a synthetic face , 2009, IUI.

[35]  F. Wallhoff,et al.  The Facial Expressions and Emotions Database Homepage (FEEDTUM) , 2005 .

[36]  R. Plutchik The measurement of emotions , 1997, Acta Neuropsychiatrica.

[37]  Wenyuan Bi,et al.  Distributing expressional faces in 2-D emotional space , 2007, CIVR '07.

[38]  Hatice Gunes,et al.  Static vs. dynamic modeling of human nonverbal behavior from multiple cues and modalities , 2009, ICMI-MLMI '09.

[39]  Tieniu Tan,et al.  Affective Information Processing , 2008 .

[40]  Björn W. Schuller,et al.  A multidimensional dynamic time warping algorithm for efficient multimodal fusion of asynchronous data streams , 2009, Neurocomputing.

[41]  Ananth N. Iyer,et al.  Emotion Detection From Infant Facial Expressions And Cries , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.

[42]  Thomas S. Huang,et al.  Emotion Recognition Based on Multimodal Information , 2009, Affective Information Processing.

[43]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques, 3rd Edition , 1999 .

[44]  Erik Cambria,et al.  Sentic Computing: Exploitation of Common Sense for the Development of Emotion-Sensitive Systems , 2009, COST 2102 Training School.

[45]  Mann Oo. Hay Emotion recognition in human-computer interaction , 2012 .

[46]  Ashish Kapoor,et al.  Automatic prediction of frustration , 2007, Int. J. Hum. Comput. Stud..

[47]  Ludmila I. Kuncheva,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2004 .

[48]  Shrikanth S. Narayanan,et al.  The Vera am Mittag German audio-visual emotional speech database , 2008, 2008 IEEE International Conference on Multimedia and Expo.

[49]  Chuan-Yu Chang,et al.  Emotion recognition with consideration of facial expression and physiological signals , 2009, 2009 IEEE Symposium on Computational Intelligence in Bioinformatics and Computational Biology.

[50]  Jennifer S. Beer,et al.  Facial expression of emotion. , 2003 .

[51]  Hatice Gunes,et al.  Bi-modal emotion recognition from expressive face and body gestures , 2007, J. Netw. Comput. Appl..

[52]  Cynthia Whissell,et al.  THE DICTIONARY OF AFFECT IN LANGUAGE , 1989 .