Cerebral processing of emotional prosody—influence of acoustic parameters and arousal

The human brain has a preference for processing of emotionally salient stimuli. In the auditory modality, emotional prosody can induce such involuntary biasing of processing resources. To investigate the neural correlates underlying automatic processing of emotional information in the voice, words spoken in neutral, happy, erotic, angry, and fearful prosody were presented in a passive-listening functional magnetic resonance imaging (fMRI) experiment. Hemodynamic responses in right mid superior temporal gyrus (STG) were significantly stronger for all emotional than for neutral intonations. To disentangle the contribution of basic acoustic features and emotional arousal to this activation, the relation between event-related responses and these parameters was evaluated by means of regression analyses. A significant linear dependency between hemodynamic responses of right mid STG and mean intensity, mean fundamental frequency, variability of fundamental frequency, duration, and arousal of the stimuli was observed. While none of the acoustic parameters alone explained the stronger responses of right mid STG to emotional relative to neutral prosody, this stronger responsiveness was abolished both by correcting for arousal or the conjoint effect of the acoustic parameters. In conclusion, our results demonstrate that right mid STG is sensitive to various emotions conveyed by prosody, an effect which is driven by a combination of acoustic features that express the emotional arousal in the speaker's voice.

[1]  L. Schiebinger,et al.  Commentary on Risto Naatanen (1990). The role of attention in auditory information processing as revealed by event-related potentials and other brain measures of cognitive fenctiono BBS 13s201-2888 , 1991 .

[2]  P. Morosan,et al.  Probabilistic Mapping and Volume Measurement of Human Primary Auditory Cortex , 2001, NeuroImage.

[3]  M. Erb,et al.  Effects of prosodic emotional intensity on activation of associative auditory cortex , 2006, Neuroreport.

[4]  Ulrich Schall,et al.  Functional neuroanatomy of auditory mismatch processing: an event-related fMRI study of duration-deviant oddballs , 2003, NeuroImage.

[5]  R. Bowtell,et al.  “sparse” temporal sampling in auditory fMRI , 1999, Human brain mapping.

[6]  Paul Boersma,et al.  Praat, a system for doing phonetics by computer , 2002 .

[7]  P. Vuilleumier,et al.  How brains beware: neural mechanisms of emotional attention , 2005, Trends in Cognitive Sciences.

[8]  R. Wise,et al.  Temporal lobe regions engaged during normal speech comprehension. , 2003, Brain : a journal of neurology.

[9]  P. Morosan,et al.  Human Primary Auditory Cortex: Cytoarchitectonic Subdivisions and Mapping into a Spatial Reference System , 2001, NeuroImage.

[10]  Erich Schröger,et al.  From Air Oscillations to Music and Speech: Functional Magnetic Resonance Imaging Evidence for Fine-Tuned Neural Networks in Audition , 2006, The Journal of Neuroscience.

[11]  R. Zatorre,et al.  Sensitivity to Auditory Object Features in Human Temporal Neocortex , 2004, The Journal of Neuroscience.

[12]  N. Tzourio-Mazoyer,et al.  Automated Anatomical Labeling of Activations in SPM Using a Macroscopic Anatomical Parcellation of the MNI MRI Single-Subject Brain , 2002, NeuroImage.

[13]  Dave R. M. Langers,et al.  fMRI activation in relation to sound intensity and loudness , 2007, NeuroImage.

[14]  R. C. Oldfield The assessment and analysis of handedness: the Edinburgh inventory. , 1971, Neuropsychologia.

[15]  I. Olson,et al.  The Enigmatic temporal pole: a review of findings on social and emotional processing. , 2007, Brain : a journal of neurology.

[16]  T. Griffiths,et al.  Functional Imaging of Pitch Analysis , 2003, Annals of the New York Academy of Sciences.

[17]  M. London,et al.  Dendritic computation. , 2005, Annual review of neuroscience.

[18]  Wolfgang Grodd,et al.  Mismatch responses to randomized gradient switching noise as reflected by fMRI and whole‐head magnetoencephalography , 2002, Human brain mapping.

[19]  Michael Erb,et al.  Cerebral pathways in processing of affective prosody: A dynamic causal modeling study , 2006, NeuroImage.

[20]  Martin Meyer,et al.  Spectro-temporal processing during speech perception involves left posterior auditory cortex , 2005, Neuroreport.

[21]  D. Collins,et al.  Automatic 3D Intersubject Registration of MR Volumetric Data in Standardized Talairach Space , 1994, Journal of computer assisted tomography.

[22]  Karl J. Friston,et al.  Statistical parametric maps in functional imaging: A general linear approach , 1994 .

[23]  Joseph S Gati,et al.  Overlapping neural regions for processing rapid temporal cues in speech and nonspeech signals☆ , 2003, NeuroImage.

[24]  Michael D. Hunter,et al.  Male and female voices activate distinct regions in the male brain , 2005, NeuroImage.

[25]  Joseph E LeDoux Emotional Memory: In Search of Systems and Synapses a , 1993, Annals of the New York Academy of Sciences.

[26]  Patrik Vuilleumier,et al.  Emotional facial expressions capture attention , 2001, Neurology.

[27]  R. Zatorre,et al.  Structure and function of auditory cortex: music and speech , 2002, Trends in Cognitive Sciences.

[28]  Michael Erb,et al.  Impact of task difficulty on lateralization of pitch and duration discrimination , 2005, Neuroreport.

[29]  K. Scherer,et al.  The voices of wrath: brain responses to angry prosody in meaningless speech , 2005, Nature Neuroscience.

[30]  Angela D. Friederici,et al.  Affective encoding in the speech signal and in event-related brain potentials , 2003, Speech Commun..

[31]  R. Näätänen The role of attention in auditory information processing as revealed by event-related potentials and other brain measures of cognitive function , 1990, Behavioral and Brain Sciences.

[32]  Doug J. K. Barrett,et al.  Response preferences for “what” and “where” in human non-primary auditory cortex , 2006, NeuroImage.

[33]  Peter Boesiger,et al.  Comparison of “silent” clustered and sparse temporal fMRI acquisitions in tonal and speech perception tasks , 2007, NeuroImage.

[34]  R. Desimone,et al.  Neural mechanisms of selective visual attention. , 1995, Annual review of neuroscience.

[35]  K. Scherer,et al.  Acoustic profiles in vocal emotion expression. , 1996, Journal of personality and social psychology.

[36]  J. Rauschecker,et al.  Mechanisms and streams for processing of "what" and "where" in auditory cortex. , 2000, Proceedings of the National Academy of Sciences of the United States of America.

[37]  R. Zatorre,et al.  Voice-selective areas in human auditory cortex , 2000, Nature.

[38]  Anne Lacheret,et al.  FMRI study of emotional speech comprehension. , 2006, Cerebral cortex.

[39]  Simon B. Eickhoff,et al.  A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data , 2005, NeuroImage.

[40]  Karl J. Friston,et al.  Classical and Bayesian Inference in Neuroimaging: Applications , 2002, NeuroImage.

[41]  A. R. Jennings,et al.  Analysis of the spectral envelope of sounds by the human brain , 2005, NeuroImage.

[42]  Klaus R. Scherer,et al.  Vocal communication of emotion: A review of research paradigms , 2003, Speech Commun..

[43]  R. Weisskoff,et al.  Effect of temporal autocorrelation due to physiological noise and stimulus paradigm on voxel‐level false‐positive rates in fMRI , 1998, Human brain mapping.

[44]  H. Scheich,et al.  Phonetic Perception and the Temporal Cortex , 2002, NeuroImage.

[45]  Anne-Lise Giraud,et al.  Distinct functional substrates along the right superior temporal sulcus for the processing of voices , 2004, NeuroImage.

[46]  S. Posse,et al.  Intensity coding of auditory stimuli: an fMRI study , 1998, Neuropsychologia.

[47]  Leslie G. Ungerleider,et al.  Repetition suppression of faces is modulated by emotion. , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[48]  Peter Boesiger,et al.  Silent and continuous fMRI scanning differentially modulate activation in an auditory language comprehension task , 2008, Human brain mapping.

[49]  Mikko Sams,et al.  Perceiving identical sounds as speech or non-speech modulates activity in the left posterior superior temporal sulcus , 2006, NeuroImage.

[50]  Sophie K. Scott,et al.  Human brain mechanisms for the early analysis of voices , 2006, NeuroImage.

[51]  R. Buxton,et al.  Modeling the hemodynamic response to brain activation , 2004, NeuroImage.

[52]  R. Henson Neuroimaging studies of priming , 2003, Progress in Neurobiology.

[53]  K. Scherer Vocal affect expression: a review and a model for future research. , 1986, Psychological bulletin.

[54]  Karl J. Friston,et al.  Modelling Geometric Deformations in Epi Time Series , 2022 .

[55]  M. Bradley,et al.  Measuring emotion: the Self-Assessment Manikin and the Semantic Differential. , 1994, Journal of behavior therapy and experimental psychiatry.

[56]  Angela D Friederici,et al.  � Human Brain Mapping 24:11–20(2005) � Voice Perception: Sex, Pitch, and the Right Hemisphere , 2022 .