Congruence of happy and sad emotion in music and faces modifies cortical audiovisual activation

BACKGROUND The powerful emotion inducing properties of music are well-known, yet music may convey differing emotional responses depending on environmental factors. We hypothesized that neural mechanisms involved in listening to music may differ when presented together with visual stimuli that conveyed the same emotion as the music when compared to visual stimuli with incongruent emotional content. METHODS We designed this study to determine the effect of auditory (happy and sad instrumental music) and visual stimuli (happy and sad faces) congruent or incongruent for emotional content on audiovisual processing using fMRI blood oxygenation level-dependent (BOLD) signal contrast. The experiment was conducted in the context of a conventional block-design experiment. A block consisted of three emotional ON periods, music alone (happy or sad music), face alone (happy or sad faces), and music combined with faces where the music excerpt was played while presenting either congruent emotional faces or incongruent emotional faces. RESULTS We found activity in the superior temporal gyrus (STG) and fusiform gyrus (FG) to be differentially modulated by music and faces depending on the congruence of emotional content. There was a greater BOLD response in STG when the emotion signaled by the music and faces was congruent. Furthermore, the magnitude of these changes differed for happy congruence and sad congruence, i.e., the activation of STG when happy music was presented with happy faces was greater than the activation seen when sad music was presented with sad faces. In contrast, incongruent stimuli diminished the BOLD response in STG and elicited greater signal change in bilateral FG. Behavioral testing supplemented these findings by showing that subject ratings of emotion in faces were influenced by emotion in music. When presented with happy music, happy faces were rated as more happy (p=0.051) and sad faces were rated as less sad (p=0.030). When presented with sad music, happy faces were rated as less happy (p=0.008) and sad faces were rated as sadder (p=0.002). INTERPRETATION Happy-sad congruence across modalities may enhance activity in auditory regions while incongruence appears to impact the perception of visual affect, leading to increased activation in face processing regions such as the FG. We suggest that greater understanding of the neural bases of happy-sad congruence across modalities can shed light on basic mechanisms of affective perception and experience and may lead to novel insights in the study of emotion regulation and therapeutic use of music.

[1]  Michael J. Constantino,et al.  Neural repetition suppression reflects fulfilled perceptual expectations , 2008 .

[2]  Scott D. Lipscomb,et al.  Perceptual judgement of the relationship between musical and visual components in film. , 1994 .

[3]  J. Sloboda,et al.  Music and Emotion , 2013 .

[4]  Jochen Kaiser,et al.  Audiovisual Functional Magnetic Resonance Imaging Adaptation Reveals Multisensory Integration Effects in Object-Related Sensory Cortices , 2010, The Journal of Neuroscience.

[5]  Talma Hendler,et al.  Feeling the real world: limbic response to music depends on related content. , 2007, Cerebral cortex.

[6]  R. Campbell,et al.  Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex , 2000, Current Biology.

[7]  J. Vroomen,et al.  Multisensory integration of emotional faces and voices in schizophrenics , 2005, Schizophrenia Research.

[8]  P. Bertelson,et al.  Multisensory integration, perception and ecological validity , 2003, Trends in Cognitive Sciences.

[9]  Jean Vroomen,et al.  Rejoinder - Bimodal emotion perception: integration across separate modalities, cross-modal perceptual grouping or perception of multimodal events? , 2000 .

[10]  G. Pourtois,et al.  Perception of Facial Expressions and Voices and of their Combination in the Human Brain , 2005, Cortex.

[11]  R. Adolphs Neural systems for recognizing emotion , 2002, Current Opinion in Neurobiology.

[12]  Hermann Ackermann,et al.  Cross-modal Interactions during Perception of Audiovisual Speech and Nonspeech Signals: An fMRI Study , 2011, Journal of Cognitive Neuroscience.

[13]  E. Bullmore,et al.  Response amplification in sensory-specific cortices during crossmodal binding. , 1999, Neuroreport.

[14]  Finn Årup Nielsen,et al.  The Brede database: a small database for functional neuroimaging , 2000 .

[15]  A. Gabrielsson Emotions in strong experiences with music. , 2001 .

[16]  I. Peretz,et al.  Exposure effects on music preference and recognition , 1998, Memory & cognition.

[17]  Steven C. R. Williams,et al.  A functional MRI study of happy and sad affective states induced by classical music , 2007, Human brain mapping.

[18]  J. Bhattacharya,et al.  Crossmodal transfer of emotion by music , 2009, Neuroscience Letters.

[19]  D. Massaro,et al.  Perceiving affect from the voice and the face , 1996, Psychonomic bulletin & review.

[20]  S. Campanella,et al.  Integrating face and voice in person perception , 2007, Trends in Cognitive Sciences.

[21]  F. Gosselin,et al.  Audio-visual integration of emotion expression , 2008, Brain Research.

[22]  Lutz Jäncke,et al.  The emotional power of music: How music enhances the feeling of affective pictures , 2006, Brain Research.

[23]  Julie Grèzes,et al.  Instrumental Music Influences Recognition of Emotional Body Language , 2009, Brain Topography.

[24]  Ruben C. Gur,et al.  Facial emotion discrimination: I. Task construction and behavioral findings in normal subjects , 1992, Psychiatry Research.

[25]  K. Grill-Spector,et al.  Repetition and the brain: neural models of stimulus-specific effects , 2006, Trends in Cognitive Sciences.

[26]  J. Vroomen,et al.  The perception of emotions by ear and by eye , 2000 .

[28]  R. Goebel,et al.  Integration of Letters and Speech Sounds in the Human Brain , 2004, Neuron.

[29]  David C. Alsop,et al.  Brain Activation during Facial Emotion Processing , 2002, NeuroImage.

[30]  M. Rigg An experiment to determine how accurately college students can interpret the intended meanings of musical compositions. , 1937 .

[31]  Rainer Goebel,et al.  Top–down task effects overrule automatic multisensory responses to letter–sound pairs in auditory association cortex , 2006, NeuroImage.

[32]  J. Sloboda,et al.  Music and emotion: Theory and research , 2001 .

[33]  H. McGurk,et al.  Hearing lips and seeing voices , 1976, Nature.

[34]  J. Haxby,et al.  The distributed human neural system for face perception , 2000, Trends in Cognitive Sciences.

[35]  Hu Cheng,et al.  Neural correlates of emotion processing in borderline personality disorder , 2009, Psychiatry Research: Neuroimaging.

[36]  Michael Erb,et al.  Impact of voice on emotional judgment of faces: An event‐related fMRI study , 2006, Human brain mapping.

[37]  R. Dolan,et al.  Crossmodal binding of fear in voice and face , 2001, Proceedings of the National Academy of Sciences of the United States of America.

[38]  S. Iversen,et al.  Detection of Audio-Visual Integration Sites in Humans by Application of Electrophysiological Criteria to the BOLD Effect , 2001, NeuroImage.

[39]  Rainer Goebel,et al.  Task‐irrelevant visual letters interact with the processing of speech sounds in heteromodal and unimodal cortex , 2008, The European journal of neuroscience.

[40]  I. Peretz,et al.  Mode and tempo relative contributions to “happy-sad” judgements in equitone melodies , 2003, Cognition & emotion.

[41]  Rainer Goebel,et al.  fMR-adaptation indicates selectivity to audiovisual content congruency in distributed clusters in human superior temporal cortex , 2010, BMC Neuroscience.