AMMON : A Speech Analysis Library for Analyzing Affect , Stress , and Mental Health on Mobile Phones

The human voice encodes a wealth of information about emotion, mood and mental state. With mobile phones this information is potentially available to a host of applications. In this paper we describe the AMMON (Affective and Mentalhealth MONitor) library, a low footprint C library designed for widely available phones. The library incorporates both core features for emotion recognition (from the Interspeech 2009 emotion recognition challenge), and the most important features for mental health analysis (glottal timing features). To comfortably run the library on feature phones (the most widely-used class of phones today), we implemented most of the routines in fixed-point arithmetic, and minimized computational and memory footprint. While there are still floating-point routines to be revised in fixed-point, on identical test data, emotion and mental stress classification accuracy was indistinguishable from a state-of-the-art reference system running on a PC.

[1]  R. Cowie,et al.  The description of naturally occurring emotional speech , 2003 .

[2]  Rosalind W. Picard,et al.  A computational model for the automatic recognition of affect in speech , 2004 .

[3]  J. Peifer,et al.  Comparing objective feature statistics of speech for classifying clinical depression , 2004, The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[4]  John H. L. Hansen,et al.  Speech Under Stress: Analysis, Modeling and Recognition , 2007, Speaker Classification.

[5]  Ricardo Gutierrez-Osuna,et al.  Using Heart Rate Monitors to Detect Mental Stress , 2009, 2009 Sixth International Workshop on Wearable and Implantable Body Sensor Networks.

[6]  Cecilia Mascolo,et al.  EmotionSense: a mobile phones based adaptive platform for experimental social psychology research , 2010, UbiComp.