Investigating the Impact of Sound Angular Position on the Listener Affective State

Emotion recognition from sound signals represents an emerging field of recent research. Although many existing works focus on emotion recognition from music, there seems to be a relative scarcity of research on emotion recognition from general sounds. One of the key characteristics of sound events is the sound source spatial position, i.e. the location of the source relatively to the acoustic receiver. Existing studies that aim to investigate the relation of the latter source placement and the elicited emotions are limited to distance, front and back spatial localization and/or specific emotional categories. In this paper we analytically investigate the effect of the source angular position on the listener's emotional state, modeled in the well-established valence/arousal affective space. Towards this aim, we have developed an annotated sound events dataset using binaural processed versions of the available International Affective Digitized Sound (IADS) sound events library. All subjective affective annotations were obtained using the Self Assessment Manikin (SAM) approach. Preliminary results obtained by processing these annotation scores are likely to indicate a systematic change in the listener affective state as the sound source angular position changes. This trend is more obvious when the sound source is located outside of the visible field of the listener.

[1]  Björn W. Schuller,et al.  Automatic recognition of emotion evoked by general sound events , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[2]  Andreas Floros,et al.  Affective acoustic ecology: towards emotionally enhanced sound events , 2012, Audio Mostly Conference.

[3]  Pratik Shah,et al.  Calibration and 3-D sound reproduction in the Immersive Audio Environment , 2011, 2011 IEEE International Conference on Multimedia and Expo.

[4]  Sanghoon Jun,et al.  A fuzzy inference-based music emotion recognition system , 2008 .

[5]  P. White Appraisal Theory , 2015 .

[6]  Paul F. M. J. Verschure,et al.  Live Soundscape Composition Based on Synthetic Emotions , 2003, IEEE Multim..

[7]  A. Ortony,et al.  What's basic about basic emotions? , 1990, Psychological review.

[8]  M. P. Bryden,et al.  A left-ear advantage for identifying the emotional quality of tonal sequences , 1982, Neuropsychologia.

[9]  K. MacDorman,et al.  Automatic Emotion Prediction of Song Excerpts: Index Construction, Algorithm Design, and Empirical Comparison , 2007 .

[10]  Nikos Moustakas,et al.  Interactive Audio Realities: An Augmented / Mixed Reality Audio Game Prototype , 2011 .

[11]  Andreas Floros,et al.  Stereo Goes Mobile: Spatial Enhancement for Short-distance Loudspeaker Setups , 2012, 2012 Eighth International Conference on Intelligent Information Hiding and Multimedia Signal Processing.

[12]  Emiliano Lorini,et al.  Computational Modeling of Emotion: Toward Improving the Inter- and Intradisciplinary Exchange , 2013, IEEE Transactions on Affective Computing.

[13]  Ana Tajadura-Jiménez,et al.  When room size matters: acoustic influences on emotional responses to sounds. , 2010, Emotion.

[14]  R. Adolphs Neural systems for recognizing emotion , 2002, Current Opinion in Neurobiology.

[15]  M. Bradley,et al.  Affective reactions to acoustic stimuli. , 2000, Psychophysiology.

[16]  J. Russell,et al.  The circumplex model of affect: An integrative approach to affective neuroscience, cognitive development, and psychopathology , 2005, Development and Psychopathology.

[17]  Rigas Kotsakis,et al.  Sound events and emotions: Investigating the relation of rhythmic characteristics and arousal , 2013, IISA 2013.

[18]  Raine Kajastila,et al.  Localization Cues Affect Emotional Judgments – Results from a User Study on Scary Sound , 2009 .

[19]  P. Laukka,et al.  Communication of emotions in vocal expression and music performance: different channels, same code? , 2003, Psychological bulletin.

[20]  William W. Gaver What in the World Do We Hear? An Ecological Approach to Auditory Event Perception , 1993 .

[21]  John Mourjopoulos,et al.  Binaural extension and performance of single-channel spectral subtraction dereverberation algorithms , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[22]  Raimund Schatz,et al.  The Impact of Test Duration on User Fatigue and Reliability of Subjective Quality Ratings , 2012 .

[23]  Patrick M. Zurek,et al.  Microphone-array hearing aids with binaural output .I. Fixed-processing systems , 1997, IEEE Trans. Speech Audio Process..

[24]  Craig A. Smith,et al.  Patterns of cognitive appraisal in emotion. , 1985, Journal of personality and social psychology.

[25]  Yi-Hsuan Yang,et al.  A Regression Approach to Music Emotion Recognition , 2008, IEEE Transactions on Audio, Speech, and Language Processing.

[26]  P. Ekman An argument for basic emotions , 1992 .

[27]  A Carmon,et al.  Ear asymmetry in perception of emotional non-verbal stimuli. , 1973, Acta psychologica.

[28]  M. Yik,et al.  A circumplex model of affect and its relation to personality : a five-language study , 1999 .

[29]  K. Scherer,et al.  On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common , 2013, Front. Psychol..

[30]  Guojun Lu,et al.  Music emotion annotation by machine learning , 2008, 2008 IEEE 10th Workshop on Multimedia Signal Processing.

[31]  Martin Ebner,et al.  Emotion Detection: Application of the Valence Arousal Space for Rapid Biological Usability Testing to Enhance Universal Access , 2009, HCI.

[32]  M. Bradley,et al.  Measuring emotion: the Self-Assessment Manikin and the Semantic Differential. , 1994, Journal of behavior therapy and experimental psychiatry.

[33]  John Mourjopoulos,et al.  Extracting Sound-Source-Distance Information from Binaural Signals , 2013 .

[34]  S. Koelsch Towards a neural basis of music-evoked emotions , 2010, Trends in Cognitive Sciences.

[35]  Ana Tajadura-Jiménez,et al.  Embodied auditory perception: the emotional impact of approaching and receding sound sources. , 2010, Emotion.

[36]  Gregory H. Wakefield,et al.  Moving Sound Source Synthesis for Binaural Electroacoustic Music Using Interpolated Head-Related Transfer Functions (HRTFs) , 2001, Computer Music Journal.

[37]  Shiliang Zhang,et al.  Affective MTV analysis based on arousal and valence features , 2008, 2008 IEEE International Conference on Multimedia and Expo.

[38]  Fredrik Hagman Emotional response to sound. Influence of spatial determinants. , 2010 .

[39]  M. Marcell,et al.  Identifying, rating, and remembering environmental sound events , 2007, Behavior research methods.

[40]  Jens Blauert,et al.  Spatial Hearing-Revised Edition : The Psychophysics of Human Sound Localization , 2017 .

[41]  K. Hevner Experimental studies of the elements of expression in music , 1936 .

[42]  Mark Grimshaw,et al.  A climate of fear: considerations for designing a virtual acoustic ecology of fear , 2011, AM '11.

[43]  K. Kroschel,et al.  Evaluation of natural emotions using self assessment manikins , 2005, IEEE Workshop on Automatic Speech Recognition and Understanding, 2005..

[44]  Tao Li,et al.  Detecting emotion in music , 2003, ISMIR.

[45]  Bill Gardner,et al.  HRTF Measurements of a KEMAR Dummy-Head Microphone , 1994 .

[46]  Peter Vary,et al.  Binaural dereverberation based on a dual-channel Wiener filter with optimized noise field coherence , 2010, 2010 IEEE International Conference on Acoustics, Speech and Signal Processing.