EmoVoice - A Framework for Online Recognition of Emotions from Voice

We present EmoVoice, a framework for emotional speech corpus and classifier creation and for offline as well as real-time online speech emotion recognition. The framework is intended to be used by non-experts and therefore comes with an interface to create an own personal or application specific emotion recogniser. Furthermore, we describe some applications and prototypes that already use our framework to track online emotional user states from voice information.

[1]  Ian Witten,et al.  Data Mining , 2000 .

[2]  Marc Cavazza,et al.  Affective Interactive Narrative in the CALLAS Project , 2007, International Conference on Virtual Storytelling.

[3]  Russell Beale,et al.  Affect and Emotion in Human-Computer Interaction, From Theory to Applications , 2008, Affect and Emotion in Human-Computer Interaction.

[4]  Christian Martyn Jones,et al.  Acoustic Emotion Recognition for Affective Computer Gaming , 2008, Affect and Emotion in Human-Computer Interaction.

[5]  Matthias Rehm,et al.  Dancing the night away: controlling a virtual karaoke dancer by multimodal expressive cues , 2008, AAMAS.

[6]  Elmar Nöth,et al.  “You Stupid Tin Box” - Children Interacting with the AIBO Robot: A Cross-linguistic Emotional Speech Corpus , 2004, LREC.

[7]  Emiel Krahmer,et al.  Real vs. acted emotional speech , 2006, INTERSPEECH.

[8]  Gernot A. Fink Developing HMM-Based Recognizers with ESMERALDA , 1999, TSD.

[9]  Johannes Wagner,et al.  Integrating information from speech and physiological signals to achieve emotional sensitivity , 2005, INTERSPEECH.

[10]  Johannes Wagner,et al.  Automatic Recognition of Emotions from Speech: A Review of the Literature and Recommendations for Practical Realisation , 2008, Affect and Emotion in Human-Computer Interaction.

[11]  F. Burkhardt,et al.  An Emotion-Aware Voice Portal , 2005 .

[12]  Britta Wrede,et al.  Playing a different imitation game: Interaction with an Empathic Android Robot , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[13]  Pierre-Yves Oudeyer,et al.  The production and recognition of emotions in speech: features and algorithms , 2003, Int. J. Hum. Comput. Stud..

[14]  Marc Cavazza,et al.  An Emotionally Responsive AR Art Installation , 2007, ISMAR 2007.

[15]  Astrid Paeschke,et al.  A database of German emotional speech , 2005, INTERSPEECH.

[16]  Elisabeth André,et al.  Improving Automatic Emotion Recognition from Speech via Gender Differentiaion , 2006, LREC.

[17]  Patrick Brézillon,et al.  Lecture Notes in Artificial Intelligence , 1999 .

[18]  Florian Schiel,et al.  The SmartKom Multimodal Corpus at BAS , 2002, LREC.

[19]  Patrice Torguet,et al.  Virtual Storytelling Using Virtual Reality Technologies for Storytelling , 2001, Lecture Notes in Computer Science.

[20]  Björn W. Schuller,et al.  Towards More Reality in the Recognition of Emotional Speech , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[21]  Christian Martyn Jones,et al.  Affective Human-Robotic Interaction , 2008, Affect and Emotion in Human-Computer Interaction.

[22]  P. Boersma Praat : doing phonetics by computer (version 4.4.24) , 2006 .

[23]  E. Velten A laboratory task for induction of mood states. , 1968, Behaviour research and therapy.

[24]  Björn Schuller,et al.  Effects of In-Car Noise-Conditions on the Recognition of Emotion within Speech , 2007 .

[25]  Joel R. Tetreault,et al.  Using system and user performance features to improve emotion detection in spoken tutoring dialogs , 2006, INTERSPEECH.

[26]  Catherine Pelachaud,et al.  From Greta's mind to her face: modelling the dynamics of affective states in a conversational embodied agent , 2003, Int. J. Hum. Comput. Stud..

[27]  Oudeyer Pierre-Yves,et al.  The production and recognition of emotions in speech: features and algorithms , 2003 .

[28]  Elisabeth André,et al.  Comparing Feature Sets for Acted and Spontaneous Speech in View of Automatic Emotion Recognition , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[29]  Paul Boersma,et al.  Praat, a system for doing phonetics by computer , 2002 .

[30]  Kornel Laskowski,et al.  Combining Efforts for Improving Automatic Classification of Emotional User States , 2006 .

[31]  Gernot A. Finkco Developing HMM-Based Recognizers with ESMERALDA , 1999 .