Ecologically valid long-term mood monitoring of individuals with bipolar disorder using speech

Speech patterns are modulated by the emotional and neurophysiological state of the speaker. There exists a growing body of work that computationally examines this modulation in patients suffering from depression, autism, and post-traumatic stress disorder. However, the majority of the work in this area focuses on the analysis of structured speech collected in controlled environments. Here we expand on the existing literature by examining bipolar disorder (BP). BP is characterized by mood transitions, varying from a healthy euthymic state to states characterized by mania or depression. The speech patterns associated with these mood states provide a unique opportunity to study the modulations characteristic of mood variation. We describe methodology to collect unstructured speech continuously and unobtrusively via the recording of day-to-day cellular phone conversations. Our pilot investigation suggests that manic and depressive mood states can be recognized from this speech data, providing new insight into the feasibility of unobtrusive, unstructured, and continuous speech-based wellness monitoring for individuals with BP.

[1]  R. C. Young,et al.  A Rating Scale for Mania: Reliability, Validity and Sensitivity , 1978, British Journal of Psychiatry.

[2]  M. Hamilton The Hamilton Rating Scale for Depression , 1986 .

[3]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[4]  William Stafford Noble,et al.  Support vector machine , 2013 .

[5]  Alan D. Lopez,et al.  Global and regional burden of disease and risk factors, 2001: systematic analysis of population health data , 2006, The Lancet.

[6]  R. Dardennes,et al.  [The costs of bipolar disorder]. , 2006, L'Encephale.

[7]  J. Mundt,et al.  Voice acoustic measures of depression severity and treatment response collected via interactive voice response (IVR) technology , 2007, Journal of Neurolinguistics.

[8]  F. Goodwin,et al.  Manic-Depressive Illness. Bipolar Disorders and Recurrent Depression. Oxford (Oxford University Press) 2007. , 2007 .

[9]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[10]  Elliot Moore,et al.  Critical Analysis of the Impact of Glottal Features in the Classification of Clinical Depression in Speech , 2008, IEEE Transactions on Biomedical Engineering.

[11]  Matthew S. Goodwin,et al.  Exploring speech therapy games with children on the autism spectrum , 2009, INTERSPEECH.

[12]  Fernando De la Torre,et al.  Detecting depression from facial actions and vocal prosody , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[13]  D K Oller,et al.  Automated vocal analysis of naturalistic recordings from children with autism, language delay, and typical development , 2010, Proceedings of the National Academy of Sciences.

[14]  Björn Schuller,et al.  Opensmile: the munich versatile and fast open-source audio feature extractor , 2010, ACM Multimedia.

[15]  S. Warren,et al.  What Automated Vocal Analysis Reveals About the Vocal Production and Language Learning Environment of Young Children with Autism , 2010, Journal of autism and developmental disorders.

[16]  Emily Tucker Prud'hommeaux,et al.  Computational prosodic markers for autism. , 2010, Autism : the international journal of research and practice.

[17]  M. McInnis,et al.  Intermediate: cognitive phenotypes in bipolar disorder. , 2010, Journal of affective disorders.

[18]  Alberto Del Bimbo,et al.  Proceedings of the international conference on Multimedia , 2010 .

[19]  E. V. D. Broek,et al.  Telling the Story and Re-Living the Past: How Speech Analysis Can Reveal Emotions in Post-traumatic Stress Disorder (PTSD) Patients , 2010 .

[20]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[21]  S. Tokuno,et al.  Usage of emotion recognition in military health care , 2011, 2011 Defense Science Research Conference and Expo (DSR).

[22]  Ton Dijkstra,et al.  Towards an Artificial Therapy Assistant - Measuring Excessive Stress from Speech , 2011, HEALTHINF.

[23]  Shrikanth S. Narayanan,et al.  Spontaneous-Speech Acoustic-Prosodic Features of Children with Autism and the Interacting Psychologist , 2012, INTERSPEECH.

[24]  Michael Wagner,et al.  From Joyous to Clinically Depressed: Mood Detection Using Spontaneous Speech , 2012, FLAIRS.

[25]  Michael Wagner,et al.  Characterising depressed speech for classification , 2013, INTERSPEECH.

[26]  Thomas F. Quatieri,et al.  Classification of depression state based on articulatory precision , 2013, INTERSPEECH.

[27]  Shrikanth S. Narayanan,et al.  Analyzing the structure of parent-moderated narratives from children with ASD using an entity-based approach , 2013, INTERSPEECH.

[28]  Roland Göcke,et al.  Modeling spectral variability for the classification of depressed speech , 2013, INTERSPEECH.

[29]  Heng Wang,et al.  Depression recognition based on dynamic facial and vocal expression features using partial least square regression , 2013, AVEC@ACM Multimedia.

[30]  Louis-Philippe Morency,et al.  Investigating voice quality as a speaker-independent indicator of depression and PTSD , 2013, INTERSPEECH.

[31]  Shan Suthaharan,et al.  Support Vector Machine , 2016 .