Cochlear Implants Special Issue Article: Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users

The present study investigated the ability of normal-hearing listeners and cochlear implant users to recognize vocal emotions. Sentences were produced by 1 male and 1 female talker according to 5 target emotions: angry, anxious, happy, sad, and neutral. Overall amplitude differences between the stimuli were either preserved or normalized. In experiment 1, vocal emotion recognition was measured in normal-hearing and cochlear implant listeners; cochlear implant subjects were tested using their clinically assigned processors. When overall amplitude cues were preserved, normal-hearing listeners achieved near-perfect performance, whereas listeners with cochlear implant recognized less than half of the target emotions. Removing the overall amplitude cues significantly worsened mean normal-hearing and cochlear implant performance. In experiment 2, vocal emotion recognition was measured in listeners with cochlear implant as a function of the number of channels (from 1 to 8) and envelope filter cutoff frequency (50 vs 400 Hz) in experimental speech processors. In experiment 3, vocal emotion recognition was measured in normal-hearing listeners as a function of the number of channels (from 1 to 16) and envelope filter cutoff frequency (50 vs 500 Hz) in acoustic cochlear implant simulations. Results from experiments 2 and 3 showed that both cochlear implant and normal-hearing performance significantly improved as the number of channels or the envelope filter cutoff frequency was increased. The results suggest that spectral, temporal, and overall amplitude cues each contribute to vocal emotion recognition. The poorer cochlear implant performance is most likely attributable to the lack of salient pitch cues and the limited functional spectral resolution.

[1]  K. Stevens,et al.  Emotions and speech: some acoustical correlates. , 1972, The Journal of the Acoustical Society of America.

[2]  William M. Rabinowitz,et al.  Better speech recognition with cochlear implants , 1991, Nature.

[3]  Qian-Jie Fu,et al.  Noise Susceptibility of Cochlear Implant Users: The Role of Spectral Resolution and Smearing , 2005, Journal of the Association for Research in Otolaryngology.

[4]  K. Peters Emotion perception in speech: Discrimination, identification, and the effects of talker and sentence variability , 2006 .

[5]  D. D. Greenwood A cochlear frequency-position function for several species--29 years later. , 1990, The Journal of the Acoustical Society of America.

[6]  R V Shannon,et al.  Speech Recognition with Primarily Temporal Cues , 1995, Science.

[7]  Iain R. Murray,et al.  Toward the simulation of emotion in synthetic speech: a review of the literature on human vocal emotion. , 1993, The Journal of the Acoustical Society of America.

[8]  F. Zeng,et al.  Speaker recognition with temporal cues in acoustic and electric hearing. , 2005, The Journal of the Acoustical Society of America.

[9]  Klaus R. Scherer,et al.  Vocal communication of emotion: A review of research paradigms , 2003, Speech Commun..

[10]  J. Galvin,et al.  The Role of Spectral and Temporal Cues in Voice Gender Discrimination by Normal-Hearing Listeners and Cochlear Implant Users , 2004, Journal of the Association for Research in Otolaryngology.

[11]  David House,et al.  Perception and production of mood in speech by cochlear implant users , 1994, ICSLP.

[12]  Efrat A. Schorr Social and Emotional Functioning of Children with Cochlear Implants , 2005 .

[13]  R. Aslin,et al.  Preference for infant-directed speech in the first month after birth. , 1990, Child development.

[14]  K. Scherer,et al.  Acoustic profiles in vocal emotion expression. , 1996, Journal of personality and social psychology.

[15]  N. Cohen,et al.  Cochlear Implants , 2000 .

[16]  R V Shannon,et al.  Effects of amplitude nonlinearity on phoneme recognition by cochlear implant users and normal-hearing listeners. , 1998, The Journal of the Acoustical Society of America.

[17]  Frank Dellaert,et al.  Recognizing emotion in speech , 1996, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96.

[18]  Julio González,et al.  Gender and speaker identification as a function of the number of channels in spectrally reduced speech. , 2005, The Journal of the Acoustical Society of America.

[19]  Q J Fu,et al.  Effects of noise and spectral resolution on vowel and consonant recognition: acoustic and electric hearing. , 1998, The Journal of the Acoustical Society of America.

[20]  B C Moore,et al.  Perceptual consequences of cochlear hearing loss and their implications for the design of hearing aids. , 1996, Ear and hearing.

[21]  Valery A. Petrushin,et al.  Emotion recognition in speech signal: experimental study, development, and application , 2000, INTERSPEECH.

[22]  Shrikanth S. Narayanan,et al.  Toward detecting emotions in spoken dialogs , 2005, IEEE Transactions on Speech and Audio Processing.

[23]  A. Fernald,et al.  Intonation and communicative intent in mothers' speech to infants: is the melody the message? , 1989, Child development.

[24]  Qian-Jie Fu,et al.  Voice gender identification by cochlear implant users: the role of spectral and temporal resolution. , 2005, The Journal of the Acoustical Society of America.

[25]  L. Trainor,et al.  Is Infant-Directed Speech Prosody a Result of the Vocal Expression of Emotion? , 2000, Psychological science.

[26]  Auditory, visual and auditory-visual identification of emotions by hearing and hearing-impaired adolescents. , 1993, British journal of audiology.

[27]  Emotion perception in pre-kindergarten school children at Central Institute for the Deaf , 2005 .

[28]  R. Shannon,et al.  Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants. , 2001, The Journal of the Acoustical Society of America.

[29]  George N. Votsis,et al.  Emotion recognition in human-computer interaction , 2001, IEEE Signal Process. Mag..

[30]  R V Shannon,et al.  A computer interface for psychophysical and speech research with the Nucleus cochlear implant. , 1990, The Journal of the Acoustical Society of America.

[31]  Zhigang Deng,et al.  An acoustic study of emotions expressed in speech , 2004, INTERSPEECH.

[32]  R. J. Lickley,et al.  Proceedings of the International Conference on Spoken Language Processing. , 1992 .