Representations of modality-general valence for videos and music derived from fMRI data

Abstract This study tested for neural representations of valence that are shared across visual and auditory modalities referred to as modality‐general representations. On a given trial participants made either affective or semantic judgments of short silent videos or music samples. For each modality valence was manipulated at three levels, positive, neutral, and negative, while controlling for the level of arousal. Whole‐brain crossmodal identification of affect indicated the presence of modality‐general valence representations that distinguished 1) positive from negative trials (signed valence) and 2) valenced from non‐valenced trials (unsigned valence). These results generalized across the two tasks. Brain regions that were sensitive to valence states in the same way for both modalities were identified by searchlight analysis of fMRI data by comparing the correlation of voxel responses to the same and different valence conditions across the two modalities. These analyses identified seven clusters that distinguished signed valence, unsigned valence or both. Signed valence was represented in the precuneus, unsigned valence in the bilateral medial prefrontal cortex, superior temporal sulcus (STS)/postcentral, and middle frontal gyrus (MFG) and both types were represented in the STS/MFG and thalamus. These results support the idea that modality general valence is represented in a network of several locations throughout the brain.

[1]  Philip A. Kragel,et al.  Multivariate neural biomarkers of emotional states are categorically distinct. , 2015, Social cognitive and affective neuroscience.

[2]  G. McCarthy,et al.  Dissociable prefrontal brain systems for attention and emotion , 2002, Proceedings of the National Academy of Sciences of the United States of America.

[3]  Jing Wang,et al.  Identifying Core Affect in Individuals from fMRI Responses to Dynamic Naturalistic Audiovisual Stimuli , 2016, PloS one.

[4]  T. Kircher,et al.  Supramodal Representation of Emotions , 2011, The Journal of Neuroscience.

[5]  Jong-Min Lee,et al.  Integration of cross-modal emotional information in the human brain: An fMRI study , 2010, Cortex.

[6]  L. F. Barrett,et al.  Affect as a Psychological Primitive. , 2009, Advances in experimental social psychology.

[7]  J. Cacioppo,et al.  Beyond Bipolar Conceptualizations and Measures: The Case of Attitudes and Evaluative Space , 1997, Personality and social psychology review : an official journal of the Society for Personality and Social Psychology, Inc.

[8]  G. Glover,et al.  Reflecting upon Feelings: An fMRI Study of Neural Systems Supporting the Attribution of Emotion to Self and Other , 2004, Journal of Cognitive Neuroscience.

[9]  Thomas Straube,et al.  Amygdala activation to fearful faces under attentional load , 2013, Behavioural Brain Research.

[10]  Rebecca Saxe,et al.  A Common Neural Code for Perceived and Inferred Emotion , 2014, The Journal of Neuroscience.

[11]  Thomas Straube,et al.  Effect of task conditions on brain responses to threatening faces in social phobics: An event-related functional magnetic resonance imaging study , 2004, Biological Psychiatry.

[12]  William A. Cunningham,et al.  Implicit and Explicit Evaluation: fMRI Correlates of Valence, Emotional Intensity, and Control in the Processing of Attitudes , 2004, Journal of Cognitive Neuroscience.

[13]  Michael Eickenberg,et al.  Correlations of correlations are not reliable statistics: implications for multivariate pattern analysis. , 2015, ICML 2015.

[14]  T. Eerola,et al.  A comparison of the discrete and dimensional models of emotion in music , 2011 .

[15]  Israel Liberzon,et al.  The neural correlates of intertemporal decision‐making: Contributions of subjective value, stimulus type, and trait impulsivity , 2011, Human brain mapping.

[16]  K. Luan Phan,et al.  Functional Neuroanatomy of Emotion: A Meta-Analysis of Emotion Activation Studies in PET and fMRI , 2002, NeuroImage.

[17]  Adam K. Anderson,et al.  Population coding of affect across stimuli, modalities and individuals , 2014, Nature Neuroscience.

[18]  Jaap Ham,et al.  Brightness differences influence the evaluation of affective pictures , 2013, Cognition & emotion.

[19]  Yuanqing Li,et al.  Selective Audiovisual Semantic Integration Enabled by Feature-Selective Attention , 2016, Scientific Reports.

[20]  Yi Chen,et al.  Statistical inference and multiple testing correction in classification-based multi-voxel pattern analysis (MVPA): Random permutations and cluster size control , 2011, NeuroImage.

[21]  Rainer Goebel,et al.  Information-based functional brain mapping. , 2006, Proceedings of the National Academy of Sciences of the United States of America.

[22]  Jing Wang,et al.  Examining Similarity Structure: Multidimensional Scaling and Related Approaches in Neuroimaging , 2013, Comput. Math. Methods Medicine.

[23]  M. Walton,et al.  Decision Making and Reward in Frontal Cortex , 2011, Behavioral neuroscience.

[24]  Philippe R Goldin,et al.  Effects of mindfulness-based stress reduction (MBSR) on emotion regulation in social anxiety disorder. , 2010, Emotion.

[25]  Luke J. Chang,et al.  A Sensitive and Specific Neural Signature for Picture-Induced Negative Affect , 2015, PLoS biology.

[26]  Qingyang Li,et al.  Emotional perception: Meta-analyses of face and natural scene processing , 2011, NeuroImage.

[27]  Tom M. Mitchell,et al.  Machine learning classifiers and fMRI: A tutorial overview , 2009, NeuroImage.

[28]  P. Lang International affective picture system (IAPS) : affective ratings of pictures and instruction manual , 2005 .

[29]  Petri Toiviainen,et al.  A Matlab Toolbox for Music Information Retrieval , 2007, GfKl.

[30]  Mikko Sams,et al.  Nonlinear relationship between emotional valence and brain activity: Evidence of separate negative and positive valence dimensions , 2009, Human brain mapping.

[31]  Jing Wang,et al.  Decoding the neural representation of affective states , 2012, NeuroImage.

[32]  B. Rossion,et al.  The time‐course of intermodal binding between seeing and hearing affective information , 2000, Neuroreport.

[33]  Nikolaus Kriegeskorte,et al.  Comparison of multivariate classifiers and response normalizations for pattern-information fMRI , 2010, NeuroImage.

[34]  A. Gabrielsson,et al.  The influence of musical structure on emotional expression. , 2001 .

[35]  Kevin N. Ochsner,et al.  Attention and emotion: Does rating emotion alter neural responses to amusing and sad films? , 2005, NeuroImage.

[36]  Lisa Feldman Barrett,et al.  The psychological construction of emotion , 2014 .

[37]  M. Bradley,et al.  Affective reactions to acoustic stimuli. , 2000, Psychophysiology.

[38]  Douglas H. Wedell,et al.  Context Effects on Similarity Judgments of Multidimensional Stimuli: Inferring the Structure of the Emotion Space , 1994 .

[39]  Alan C. Evans,et al.  Emotional responses to pleasant and unpleasant music correlate with activity in paralimbic brain regions , 1999, Nature Neuroscience.

[40]  R. Schultz,et al.  Superior temporal activation in response to dynamic audio-visual emotional cues , 2009, Brain and Cognition.

[41]  M. Peelen,et al.  Supramodal Representations of Perceived Emotions in the Human Brain , 2010, The Journal of Neuroscience.

[42]  Shelley E. Taylor,et al.  Asymmetrical effects of positive and negative events: the mobilization-minimization hypothesis. , 1991, Psychological bulletin.

[43]  Yuanqing Li,et al.  Crossmodal integration enhances neural representation of task-relevant features in audiovisual face perception. , 2015, Cerebral cortex.

[44]  Anjali Krishnan,et al.  Cluster-extent based thresholding in fMRI analyses: Pitfalls and recommendations , 2014, NeuroImage.

[45]  Thomas E. Nichols,et al.  Nonparametric permutation tests for functional neuroimaging: A primer with examples , 2002, Human brain mapping.

[46]  J. Russell,et al.  Independence and bipolarity in the structure of current affect. , 1998 .

[47]  Jeffrey M. Zacks,et al.  Searchlight analysis: Promise, pitfalls, and potential , 2013, NeuroImage.

[48]  K. Scherer,et al.  The World of Emotions is not Two-Dimensional , 2007, Psychological science.

[49]  Annie Pye,et al.  Evidence for a supra-modal representation of emotion from cross-modal adaptation , 2015, Cognition.

[50]  Samuel M. McClure,et al.  Separate Neural Systems Value Immediate and Delayed Monetary Rewards , 2004, Science.

[51]  Mikko Sams,et al.  Representation of perceived sound valence in the human brain , 2012, Human brain mapping.

[52]  Ajay B. Satpute,et al.  The Brain Basis of Positive and Negative Affect: Evidence from a Meta-Analysis of the Human Neuroimaging Literature. , 2016, Cerebral cortex.

[53]  Jing Wang,et al.  Representations of modality‐specific affective processing for visual and auditory stimuli derived from functional magnetic resonance imaging data , 2014, Human brain mapping.

[54]  Seong-Whan Lee,et al.  Abstract Representations of Associated Emotions in the Human Brain , 2015, The Journal of Neuroscience.

[55]  D. Watson,et al.  Toward a consensual structure of mood. , 1985, Psychological bulletin.

[56]  R. Hari,et al.  Discrete Neural Signatures of Basic Emotions. , 2016, Cerebral cortex.

[57]  M. Just,et al.  Identifying Emotions on the Basis of Neural Activation , 2013, PloS one.

[58]  Pascal Belin,et al.  Crossmodal Adaptation in Right Posterior Superior Temporal Sulcus during Face–Voice Emotional Integration , 2014, The Journal of Neuroscience.

[59]  A. Young,et al.  Task instructions modulate neural responses to fearful facial expressions , 2003, Biological Psychiatry.

[60]  Jongwan Kim,et al.  Comparison of physiological responses to affect eliciting pictures and music. , 2016, International journal of psychophysiology : official journal of the International Organization of Psychophysiology.

[61]  B. Argall,et al.  Integration of Auditory and Visual Information about Objects in Superior Temporal Sulcus , 2004, Neuron.

[62]  J. Cacioppo,et al.  The affect system has parallel and integrative processing components: Form follows function. , 1999 .

[63]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[64]  Michael Erb,et al.  Audiovisual integration of emotional signals in voice and face: An event-related fMRI study , 2007, NeuroImage.