Relating Perceptual and Feature Space Invariances in Music Emotion Recognition

It is natural for people to organize music in terms of its emo- tional associations, but while this task is a natural process for humans, quantifying it empirically proves to be a very dicult task. Consequently, no particular acoustic feature has emerged as the optimal representation for musical emotion recognition. Due to the subjective nature of emotion, determining how informative an acoustic feature domain is requires eval- uation by human subjects. In this work, we seek to perceptually evaluate two of the most commonly used features in music information retrieval: mel-frequency cepstral coecients and the chromagram. Furthermore, to identify emotion-informative feature domains, we seek to identify what musical features are most variant or invariant to changes in musical qual- ities. This information could also potentially be used to inform methods that seek to learn acoustic representations that are specifically optimized for prediction of emotion.

[1]  Youngmoo E. Kim,et al.  Learning emotion-based acoustic features with deep belief networks , 2011, 2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA).

[2]  E. Schellenberg,et al.  Effects of Musical Tempo and Mode on Arousal, Mood, and Spatial Abilities , 2002 .

[3]  Gregory D. Webster,et al.  Emotional Responses to Music: Interactive Effects of Mode, Texture, and Tempo , 2005 .

[4]  R. Thayer The biopsychology of mood and arousal , 1989 .

[5]  L. Gerken,et al.  The Development of Affective Responses to Modality and Melodic Contour , 1995 .

[6]  Youngmoo E. Kim,et al.  Modeling Musical Emotion Dynamics with Conditional Random Fields , 2011, ISMIR.

[7]  Youngmoo E. Kim,et al.  Feature selection for content-based, time-varying musical emotion regression , 2010, MIR '10.

[8]  Jeffrey J. Scott,et al.  MUSIC EMOTION RECOGNITION: A STATE OF THE ART REVIEW , 2010 .

[9]  P. Juslin,et al.  Play it again with feeling: computer feedback in musical communication of emotions. , 2006, Journal of experimental psychology. Applied.

[10]  I. Peretz,et al.  A developmental study of the affective value of tempo and mode in music , 2001, Cognition.

[11]  Youngmoo E. Kim,et al.  Prediction of Time-varying Musical Mood Distributions from Audio , 2010, ISMIR.

[12]  Douglas Eck,et al.  Learning Tags that Vary Within a Song , 2010, ISMIR.

[13]  M. Rigg Speed as a determiner of musical mood. , 1940 .

[14]  K. Hevner Experimental studies of the elements of expression in music , 1936 .

[15]  I. Peretz,et al.  Mode and tempo relative contributions to “happy-sad” judgements in equitone melodies , 2003, Cognition & emotion.

[16]  Brandon G. Morton,et al.  A Comparative Study of Collaborative vs. Traditional Musical Mood Annotation , 2011, ISMIR.

[17]  Jin Ha Lee,et al.  Crowdsourcing Music Similarity Judgments using Mechanical Turk , 2010, ISMIR.