Toward Multimodal Modeling of Emotional Expressiveness

Emotional expressiveness captures the extent to which a person tends to outwardly display their emotions through behavior. Due to the close relationship between emotional expressiveness and behavioral health, as well as the crucial role that it plays in social interaction, the ability to automatically predict emotional expressiveness stands to spur advances in science, medicine, and industry. In this paper, we explore three related research questions. First, how well can emotional expressiveness be predicted from visual, linguistic, and multimodal behavioral signals? Second, how important is each behavioral modality to the prediction of emotional expressiveness? Third, which behavioral signals are reliably related to emotional expressiveness? To answer these questions, we add highly reliable transcripts and human ratings of perceived emotional expressiveness to an existing video database and use this data to train, validate, and test predictive models. Our best model shows promising predictive performance on this dataset (RMSE=0.65, R^2=0.45, r=0.74). Multimodal models tend to perform best overall, and models trained on the linguistic modality tend to outperform models trained on the visual modality. Finally, examination of our interpretable models' coefficients reveals a number of visual and linguistic behavioral signals---such as facial action unit intensity, overall word count, and use of words related to social processes---that reliably predict emotional expressiveness.

[1]  D. A. Smith,et al.  Individual differences in dispositional expressiveness: development and validation of the Emotional Expressivity Scale. , 1994, Journal of personality and social psychology.

[2]  Robert Rosenthal,et al.  Conducting Judgment Studies , 2008 .

[3]  A. Freitas-Magalhães Facial Expression of Emotion , 2012 .

[4]  Trevor Darrell,et al.  DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition , 2013, ICML.

[5]  Terry K Koo,et al.  A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. , 2016, Journal of chiropractic medicine.

[6]  Qiang Ji,et al.  Facial Action Unit Recognition by Exploiting Their Dynamic and Semantic Relationships , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Fabio Valente,et al.  Annotation and Recognition of Personality Traits in Spoken Conversations from the AMI Meetings Corpus , 2012, INTERSPEECH.

[8]  Nicu Sebe,et al.  Facial expression recognition from video sequences: temporal and static modeling , 2003, Comput. Vis. Image Underst..

[9]  George Trigeorgis,et al.  End-to-End Multimodal Emotion Recognition Using Deep Neural Networks , 2017, IEEE Journal of Selected Topics in Signal Processing.

[10]  P. Bentler,et al.  Cutoff criteria for fit indexes in covariance structure analysis : Conventional criteria versus new alternatives , 1999 .

[11]  Alessandro Vinciarelli,et al.  A Survey of Personality Computing , 2014, IEEE Transactions on Affective Computing.

[12]  Subramanian Ramanathan,et al.  Automatic modeling of personality states in small group interactions , 2011, MM '11.

[13]  Ruben C. Arslan,et al.  formr: A study framework allowing for automated feedback generation and complex longitudinal experience-sampling studies using R , 2018, Behavior Research Methods.

[14]  Elmar Nöth,et al.  The INTERSPEECH 2012 Speaker Trait Challenge , 2012, INTERSPEECH.

[15]  Daniel M. McNeish,et al.  Thinking twice about sum scores , 2020, Behavior Research Methods.

[16]  Claude Montacié,et al.  Pitch and Intonation Contribution to Speakers' Traits Classification , 2012, INTERSPEECH.

[17]  N L Segal,et al.  Personality similarity in twins reared apart and together. , 1988, Journal of personality and social psychology.

[18]  Louis-Philippe Morency,et al.  Context-Dependent Models for Predicting and Characterizing Facial Expressiveness , 2019, AffCon@AAAI.

[19]  Tamás D. Gedeon,et al.  EmotiW 2018: Audio-Video, Student Engagement and Group-Level Affect Prediction , 2018, ICMI.

[20]  Marcello Gallucci,et al.  A conceptual and empirical examination of justifications for dichotomization. , 2009, Psychological methods.

[21]  Jessica Kay Flake,et al.  Construct Validation in Social and Personality Research , 2017 .

[22]  Ashish Kapoor,et al.  Multimodal affect recognition in learning environments , 2005, ACM Multimedia.

[23]  Daniel O. Segall,et al.  Understanding and Assessing Nonverbal Expressiveness: The Affective Communication Test. , 1980 .

[24]  M. Kenward,et al.  An Introduction to the Bootstrap , 2007 .

[25]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[26]  Rex B. Kline,et al.  Principles and Practice of Structural Equation Modeling , 1998 .

[27]  Louis-Philippe Morency,et al.  OpenFace 2.0: Facial Behavior Analysis Toolkit , 2018, 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018).

[28]  Robert R. McCrae,et al.  Bibliography for the Revised NEO Personality Inventory (NEO PI-R ™ ) and NEO Five-Factor Inventory (NEO-FFI ™ ) , 2003 .

[29]  Yuxiao Hu,et al.  Audio-Visual Spontaneous Emotion Recognition , 2007, Artifical Intelligence for Human Computing.

[30]  T. Sullivan,et al.  Observed Emotional Expressivity, Posttraumatic Stress Disorder Symptoms, and Intimate Partner Violence Perpetration Among Community Couples. , 2018, Journal of traumatic stress.

[31]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .

[32]  James J. Gross,et al.  Facets of emotional Expressivity: Three self-report factors and their correlates , 1995 .

[33]  P. Costa,et al.  NEO inventories for the NEO Personality Inventory-3 (NEO-PI-3), NEO Five-Factor Inventory-3 (NEO-FFI-3), NEO Personality Inventory-Revised (NEO PI-R) : professional manual , 2010 .

[34]  P. Costa,et al.  The five-factor theory of personality. , 2008 .

[35]  Stefan Carlsson,et al.  CNN Features Off-the-Shelf: An Astounding Baseline for Recognition , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[36]  K. Gwet Handbook of Inter-Rater Reliability: The Definitive Guide to Measuring the Extent of Agreement Among Raters , 2014 .

[37]  Gwen Littlewort,et al.  Faces of pain: automated measurement of spontaneousallfacial expressions of genuine and posed pain , 2007, ICMI '07.

[38]  Itir Onal Ertugrul,et al.  Affective facial computing: Generalizability across domains , 2019, Multimodal Behavior Analysis in the Wild.

[39]  Jeffrey F. Cohn,et al.  Sayette Group Formation Task (GFT) Spontaneous Facial Expression Database , 2017, 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017).

[40]  Tim Polzehl,et al.  Automatically Assessing Personality from Speech , 2010, 2010 IEEE Fourth International Conference on Semantic Computing.

[41]  Yves Rosseel,et al.  blavaan: Bayesian structural equation models via parameter expansion , 2015, 1511.05604.

[42]  K. McGraw,et al.  Forming inferences about some intraclass correlation coefficients. , 1996 .

[43]  Maja Pantic,et al.  Audiovisual discrimination between laughter and speech , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.

[44]  Alessandro Vinciarelli,et al.  Automatic Personality Perception: Prediction of Trait Attribution Based on Prosodic Features , 2012, IEEE Transactions on Affective Computing.

[45]  Alan S. Cowen,et al.  Mapping the Passions: Toward a High-Dimensional Taxonomy of Emotional Experience and Expression , 2019, Psychological science in the public interest : a journal of the American Psychological Society.

[46]  Ryan L. Boyd,et al.  The Development and Psychometric Properties of LIWC2015 , 2015 .

[47]  Björn W. Schuller,et al.  Categorical and dimensional affect analysis in continuous input: Current trends and future directions , 2013, Image Vis. Comput..

[48]  John M. Gottman,et al.  Responsive Listening in Long-Married Couples: A Psycholinguistic Perspective , 1999 .

[49]  Shan Li,et al.  Deep Facial Expression Recognition: A Survey , 2018, IEEE Transactions on Affective Computing.

[50]  Björn W. Schuller,et al.  Audiovisual recognition of spontaneous interest within conversations , 2007, ICMI '07.

[51]  Daniel Gatica-Perez,et al.  The YouTube Lens: Crowdsourced Personality Impressions and Audiovisual Analysis of Vlogs , 2013, IEEE Transactions on Multimedia.

[52]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[53]  Sergio Escalera,et al.  ChaLearn LAP 2016: First Round Challenge on First Impressions - Dataset and Results , 2016, ECCV Workshops.

[54]  Terrence D. Jorgensen,et al.  Adapting fit indices for Bayesian structural equation modeling: Comparison to maximum likelihood. , 2020, Psychological methods.

[55]  Kerstin Eder,et al.  Believing in BERT: Using expressive communication to enhance trust and counteract operational error in physical Human-robot interaction , 2016, 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[56]  Abhinav Dhall,et al.  Emotion recognition in the wild challenge 2013 , 2013, ICMI '13.

[57]  Aleix M. Martinez,et al.  Emotional Expressions Reconsidered: Challenges to Inferring Emotion From Human Facial Movements , 2019, Psychological science in the public interest : a journal of the American Psychological Society.

[58]  Aidan G. C. Wright,et al.  Current Directions in Personality Science and the Potential for Advances through Computing , 2014, IEEE Transactions on Affective Computing.

[59]  Tamás D. Gedeon,et al.  Video and Image based Emotion Recognition Challenges in the Wild: EmotiW 2015 , 2015, ICMI.

[60]  Laurence Devillers,et al.  Detection of real-life emotions in call centers , 2005, INTERSPEECH.

[61]  Peter Harms,et al.  An evaluation of the consequences of using short measures of the Big Five personality traits. , 2012, Journal of personality and social psychology.

[62]  L. King,et al.  Conflict over emotional expression: psychological and physical correlates. , 1990, Journal of personality and social psychology.

[63]  Zhihong Zeng,et al.  A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[64]  Laurence Devillers,et al.  Personality traits detection using a parallelized modified SFFS algorithm , 2012, INTERSPEECH.

[65]  P. Ekman,et al.  Facial action coding system: a technique for the measurement of facial movement , 1978 .