Implicit Media Tagging and Affect Prediction from RGB-D Video of Spontaneous Facial Expressions

We present a method that automatically evaluates emotional response from spontaneous facial activity. The automatic evaluation of emotional response, or affect, is a fascinating challenge with many applications. Our approach is based on the inferred activity of facial muscles over time, as automatically obtained from an RGB-D video recording of spontaneous facial activity. Our contribution is two-fold: First, we constructed a database of publicly available short video clips, which elicit a strong emotional response in a consistent manner across different individuals. Each video was tagged by its characteristic emotional response along 4 scales: Valence, Arousal, Likability and Rewatch (the desire to watch again). The second contribution is a two-step prediction method, based on learning, which was trained and tested using this database of tagged video clips. Our method was able to successfully predict the aforementioned 4 dimensional representation of affect, achieving high correlation (0.87-0.95) between the predicted scores and the affect tags. As part of the prediction algorithm we identified the period of strongest emotional response in the viewing recordings, in a method that was blind to the video clip being watched, showing high agreement between independent viewers. Finally, inspection of the relative contribution of different feature types to the prediction process revealed that temporal facets contributed more to the prediction of individual affect than to media tags.

[1]  Rongrong Ji,et al.  Video indexing and recommendation based on affective analysis of viewers , 2011, MM '11.

[2]  Ó. Gonçalves,et al.  The Emotional Movie Database (EMDB): A Self-Report and Psychophysiological Study , 2012, Applied psychophysiology and biofeedback.

[3]  Björn W. Schuller,et al.  Emotion representation, analysis and synthesis in continuous space: A survey , 2011, Face and Gesture 2011.

[4]  Qiang Ji,et al.  Implicit video emotion tagging from audiences’ facial expression , 2013, Multimedia Tools and Applications.

[5]  M. Bradley,et al.  Measuring emotion: the Self-Assessment Manikin and the Semantic Differential. , 1994, Journal of behavior therapy and experimental psychiatry.

[6]  Mohammad Soleymani,et al.  Analysis of EEG Signals and Facial Expressions for Continuous Emotion Detection , 2016, IEEE Transactions on Affective Computing.

[7]  Daniel McDuff,et al.  Event detection: Ultra large-scale clustering of facial expressions , 2015, 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[8]  Daniel Hadar,et al.  Implicit Media Tagging and Affect Prediction from video of spontaneous facial expressions, recorded with depth camera , 2017, ArXiv.

[9]  P. Ekman,et al.  Facial action coding system , 2019 .

[10]  Thierry Pun,et al.  DEAP: A Database for Emotion Analysis ;Using Physiological Signals , 2012, IEEE Transactions on Affective Computing.

[11]  P. Lang The emotion probe. Studies of motivation and attention. , 1995, The American psychologist.

[12]  Ayoub Al-Hamadi,et al.  Emotion Recognition based on 2D-3D Facial Feature Extraction from Color Image Sequences , 2010, J. Multim..

[13]  Daphna Weinshall,et al.  Automated Facial Expressions Analysis in Schizophrenia: A Continuous Dynamic Approach , 2015, MindCare.

[14]  Andrea Cavallaro,et al.  Automatic Analysis of Facial Affect: A Survey of Registration, Representation, and Recognition , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Mohammad Soleymani,et al.  Human-centered implicit tagging: Overview and perspectives , 2012, 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[16]  Romit Roy Choudhury,et al.  Your reactions suggest you liked the movie: automatic content rating via reaction sensing , 2013, UbiComp.

[17]  Rana El Kaliouby,et al.  Automatic measurement of ad preferences from facial responses gathered over the Internet , 2014, Image Vis. Comput..

[18]  Le An,et al.  Zapping Index:Using Smile to Measure Advertisement Zapping Likelihood , 2014, IEEE Transactions on Affective Computing.

[19]  Emad Barsoum,et al.  Emotion recognition in the wild from videos using images , 2016, ICMI.

[20]  Daphna Weinshall,et al.  Facial expressions and flat affect in schizophrenia, automatic analysis from depth camera data , 2016, 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI).