Attentive to Individual: A Multimodal Emotion Recognition Network with Personalized Attention Profile

A growing number of human-centered applications benefit from continuous advancements in the emotion recognition technology. Many emotion recognition algorithms have been designed to model multimodal behavior cues to achieve high performances. However, most of them do not consider the modulating factors of an individual’s personal attributes in his/her expressive behaviors. In this work, we propose a Personalized Attributes-Aware Attention Network (PAaAN) with a novel personalized attention mechanism to perform emotion recognition using speech and language cues. The attention profile is learned from embeddings of an individual’s profile, acoustic, and lexical behavior data. The profile embedding is derived using linguistics inquiry word count computed between the target speaker and a large set of movie scripts. Our method achieves the stateof-the-art 70.3% unweighted accuracy in a four class emotion recognition task on the IEMOCAP. Further analysis reveals that affect-related semantic categories are emphasized differently for each speaker in the corpus showing the effectiveness of our attention mechanism for personalization.

[1]  Carlos Busso,et al.  IEMOCAP: interactive emotional dyadic motion capture database , 2008, Lang. Resour. Evaluation.

[2]  Sarah E Hampson,et al.  Personality processes: mechanisms by which personality traits "get outside the skin". , 2012, Annual review of psychology.

[3]  Najim Dehak,et al.  Deep Neural Networks for Emotion Recognition Combining Audio and Transcripts , 2018, INTERSPEECH.

[4]  Reginald B. Adams,et al.  Who may frown and who should smile? Dominance, affiliation, and the display of happiness and anger , 2005 .

[5]  Margaret L. Kern,et al.  Personality, Gender, and Age in the Language of Social Media: The Open-Vocabulary Approach , 2013, PloS one.

[6]  Björn Schuller,et al.  Opensmile: the munich versatile and fast open-source audio feature extractor , 2010, ACM Multimedia.

[7]  Ryan L. Boyd,et al.  The Development and Psychometric Properties of LIWC2015 , 2015 .

[8]  Björn W. Schuller,et al.  The effect of personality trait, age, and gender on the performance of automatic speech valence recognition , 2017, 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII).

[9]  Cristian Danescu-Niculescu-Mizil,et al.  Chameleons in Imagined Conversations: A New Approach to Understanding Coordination of Linguistic Style in Dialogs , 2011, CMCL@ACL.

[10]  Raymond W. M. Ng,et al.  Multi-Modal Sequence Fusion via Recursive Attention for Emotion Recognition , 2018, CoNLL.

[11]  J. Silk,et al.  The Role of the Family Context in the Development of Emotion Regulation. , 2007, Social development.

[12]  Russell P. Guay,et al.  Personality, values, and motivation , 2009 .

[13]  Panayiotis G. Georgiou,et al.  Behavioral Signal Processing: Deriving Human Behavioral Informatics From Speech and Language , 2013, Proceedings of the IEEE.

[14]  David Matsumoto,et al.  Are Cultural Differences in Emotion Regulation Mediated by Personality Traits? , 2006 .

[15]  Igor Bisio,et al.  Gender-Driven Emotion Recognition Through Speech Signals For Ambient Intelligence Applications , 2013, IEEE Transactions on Emerging Topics in Computing.

[16]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[17]  Thorsten Joachims,et al.  Improving Recommender Systems Beyond the Algorithm , 2018, ArXiv.

[18]  A. Kring,et al.  Sex differences in emotion: expression, experience, and physiology. , 1998, Journal of personality and social psychology.

[19]  Yue Gao,et al.  Personality-Aware Personalized Emotion Recognition from Physiological Signals , 2018, IJCAI.

[20]  P. Ekman An argument for basic emotions , 1992 .

[21]  Matteo Bianchi,et al.  A Human–Robot Interaction Perspective on Assistive and Rehabilitation Robotics , 2017, Front. Neurorobot..

[22]  Lun-Wei Ku,et al.  EmotionLines: An Emotion Corpus of Multi-Party Conversations , 2018, LREC.

[23]  Erik Cambria,et al.  Convolutional MKL Based Multimodal Emotion Recognition and Sentiment Analysis , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[24]  Stefanie Rukavina,et al.  Affective Computing and the Impact of Gender and Age , 2016, PloS one.

[25]  T. Shackelford,et al.  The SAGE Handbook of Personality and Individual Differences: Volume III: Applications of Personality and IndividualDifferences , 2018 .