Automatic Recognition of Affective Laughter in Spontaneous Dyadic Interactions from Audiovisual Signals

Laughter is a highly spontaneous behavior that frequently occurs during social interactions. It serves as an expressive-communicative social signal which conveys a large spectrum of affect display. Even though many studies have been performed on the automatic recognition of laughter -- or emotion -- from audiovisual signals, very little is known about the automatic recognition of emotion conveyed by laughter. In this contribution, we provide insights on emotional laughter by extensive evaluations carried out on a corpus of dyadic spontaneous interactions, annotated with dimensional labels of emotion (arousal and valence). We evaluate, by automatic recognition experiments and correlation based analysis, how different categories of laughter, such as unvoiced laughter, voiced laughter, speech laughter, and speech (non-laughter) can be differentiated from audiovisual features, and to which extent they might convey different emotions. Results show that voiced laughter performed best in the automatic recognition of arousal and valence for both audio and visual features. The context of production is further analysed and results show that, acted and spontaneous expressions of laughter produced by a same person can be differentiated from audiovisual signals, and multilingual induced expressions can be differentiated from those produced during interactions.

[1]  Thierry Dutoit,et al.  The AVLaughterCycle Database , 2010, LREC.

[2]  Merlin Suarez,et al.  Building a Multimodal Laughter Database for Emotion Recognition , 2012, LREC.

[3]  R. Stibbard AUTOMATED EXTRACTION OF ToBI ANNOTATION DATA FROM THE READING / LEEDS EMOTIONAL SPEECH CORPUS , 2000 .

[4]  P. Ekman,et al.  The Duchenne smile: emotional expression and brain physiology. II. , 1990, Journal of personality and social psychology.

[5]  Radoslaw Niewiadomski,et al.  MMLI: Multimodal Multiperson Corpus of Laughter in Interaction , 2013, HBU.

[6]  W. Chafe The Importance of Not Being Earnest , 2007 .

[7]  J. Bachorowski,et al.  The acoustic features of human laughter. , 2001, The Journal of the Acoustical Society of America.

[8]  Fabien Ringeval,et al.  Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge , 2014, AVEC@ACM Multimedia.

[9]  Margaret McRorie,et al.  The Belfast Induced Natural Emotion Database , 2012, IEEE Transactions on Affective Computing.

[10]  William Curran,et al.  The Belfast storytelling database: A spontaneous social interaction database with laughter focused annotation , 2015, 2015 International Conference on Affective Computing and Intelligent Interaction (ACII).

[11]  Wallace L. Chafe,et al.  The Importance of Not Being Earnest – The Feeling behind Laughter and Humor , 2007, Phonetica.

[12]  Jean-Philippe Thiran,et al.  Prediction of asynchronous dimensional emotion ratings from audiovisual and physiological data , 2015, Pattern Recognit. Lett..

[13]  Björn W. Schuller,et al.  The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing , 2016, IEEE Transactions on Affective Computing.

[14]  Björn Schuller,et al.  On Laughter and Speech-Laugh, Based on Observations of Child-Robot Interaction , 2019, ArXiv.

[15]  R A Martin,et al.  Humor, laughter, and physical health: methodological issues and research findings. , 2001, Psychological bulletin.

[16]  Jürgen Trouvain,et al.  Laughter annotations in conversational speech corpora - possibilities and limitations for phonetic analysis , 2012 .

[17]  Maja Pantic,et al.  Audiovisual Discrimination Between Speech and Laughter: Why and When Visual Information Might Help , 2011, IEEE Transactions on Multimedia.

[18]  Björn Schuller,et al.  The Computational Paralinguistics Challenge , 2012 .

[19]  P. Ekman,et al.  What the face reveals : basic and applied studies of spontaneous expression using the facial action coding system (FACS) , 2005 .

[20]  J. Bachorowski,et al.  Not All Laughs are Alike: Voiced but Not Unvoiced Laughter Readily Elicits Positive Affect , 2001, Psychological science.

[21]  Fabien Ringeval,et al.  Emotion Recognition in the Wild: Incorporating Voice and Lip Activity in Multimodal Decision-Level Fusion , 2014, ICMI.

[22]  A. Fogel,et al.  The integration of laughter and speech in vocal communication: a dynamic systems perspective. , 1999, Journal of speech, language, and hearing research : JSLHR.

[23]  Fabien Ringeval,et al.  Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions , 2013, 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[24]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[25]  Maja Pantic,et al.  Audiovisual Detection of Laughter in Human-Machine Interaction , 2013, 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction.

[26]  Phillip J. Glenn Laughter in Interaction , 2003 .

[27]  Joshua Foer,et al.  Laughter: A Scientific Investigation , 2001, The Yale Journal of Biology and Medicine.

[28]  J. Orozco,et al.  Spatio-temporal reasoning for reliable facial expression interpretation , 2007 .

[29]  C. Darwin The Expression of the Emotions in Man and Animals , .

[30]  P. Ekman,et al.  The expressive pattern of laughter , 2001 .

[31]  Peter Robinson,et al.  Cross-dataset learning and person-specific normalisation for automatic Action Unit detection , 2015, 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[32]  Marc Schröder,et al.  Experimental study of affect bursts , 2003, Speech Commun..

[33]  Björn W. Schuller,et al.  Speech-Based Non-Prototypical Affect Recognition for Child-Robot Interaction in Reverberated Environments , 2011, INTERSPEECH.

[34]  Kornel Laskowski,et al.  Analysis of the occurrence of laughter in meetings , 2007, INTERSPEECH.

[35]  Maja Pantic,et al.  AVEC 2014 – The Three Dimensional Affect and Depression Challenge , 2014 .

[36]  Thierry Dutoit,et al.  AVLaughterCycle : Enabling a virtual agent to join in laughing with a conversational partner using a similarity-driven audiovisual laughter animation (Original Paper) , 2010 .

[37]  Shrikanth S. Narayanan,et al.  Robust Unsupervised Arousal Rating:A Rule-Based Framework withKnowledge-Inspired Vocal Features , 2014, IEEE Transactions on Affective Computing.

[38]  Kornel Laskowski,et al.  Contrasting emotion-bearing laughter types in multiparticipant vocal activity detection for meetings , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.

[39]  Katherine B. Martin,et al.  Facial Action Coding System , 2015 .

[40]  William Curran,et al.  An Event Driven Fusion Approach for Enjoyment Recognition in Real-time , 2014, ACM Multimedia.

[41]  L. Devillers,et al.  Positive and Negative emotional states behind the laughs in spontaneous spoken dialogs , 2007 .

[42]  Maja Pantic,et al.  The MAHNOB Laughter database , 2013, Image Vis. Comput..

[43]  Li-Qun Xu,et al.  User-oriented affective video content analysis , 2001, Proceedings IEEE Workshop on Content-Based Access of Image and Video Libraries (CBAIVL 2001).

[44]  Björn W. Schuller,et al.  Recent developments in openSMILE, the munich open-source multimedia feature extractor , 2013, ACM Multimedia.

[45]  Atsuo Takanishi,et al.  Quantitative Laughter Detection, Measurement, and Classification—A Critical Survey , 2016, IEEE Reviews in Biomedical Engineering.

[46]  P. Boersma Praat : doing phonetics by computer (version 5.1.05) , 2009 .

[47]  J. Trouvain Segmenting Phonetic Units in Laughter , 2003 .

[48]  J. Trouvain Phonetic Aspects of "Speech-Laughs" , 2001 .

[49]  Maja Pantic,et al.  Audiovisual laughter detection based on temporal features , 2008, ICMI '08.

[50]  Günther Palm,et al.  Multimodal Laughter Detection in Natural Discourses , 2009, Human Centered Robot Systems, Cognition, Interaction, Technology.

[51]  C. Schaefer,et al.  Effects of Laughing, Smiling, and Howling on Mood , 2002, Psychological reports.

[52]  Hennie Brugman,et al.  Annotating Multi-media/Multi-modal Resources with ELAN , 2004, LREC.

[53]  Maja Pantic,et al.  This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING , 2022 .

[54]  Dirk Wildgruber,et al.  Differentiation of emotions in laughter at the behavioral level. , 2009, Emotion.

[55]  W. Stone,et al.  Laughter Differs in Children with Autism: An Acoustic Analysis of Laughs Produced by Children With and Without the Disorder , 2009, Journal of autism and developmental disorders.

[56]  Fabien Ringeval,et al.  AV+EC 2015: The First Affect Recognition Challenge Bridging Across Audio, Video, and Physiological Data , 2015, AVEC@ACM Multimedia.

[57]  Jay Hall,et al.  The Effects of a Normative Intervention on Group Decision-Making Performance , 1970 .

[58]  Fabien Ringeval,et al.  AVEC 2016: Depression, Mood, and Emotion Recognition Workshop and Challenge , 2016, AVEC@ACM Multimedia.

[59]  Björn W. Schuller,et al.  The Computational Paralinguistics Challenge [Social Sciences] , 2012, IEEE Signal Processing Magazine.

[60]  Tanja Schultz,et al.  Detection of Laughter-in-Interaction in Multichannel Close-Talk Microphone Recordings of Meetings , 2008, MLMI.