Music Emotion Retrieval Based on Acoustic Features

Music emotion expresses inherent and high-level states of mind and spiritual quality. In this paper, a hierarchical framework is proposed, which consists of two layers: an external layer that represents preliminary and superficial emotions and an inherent layer that represents psychic and resonant emotions. Using these two layers, a Resonance-Arousal-Valence (RAV) emotion model has been constructed. Five feature sets, including intensity, timbre, rhythm, pitch and tonality, and harmony, are extracted to represent music emotions in the RAV model. In order to effectively represent emotions with extracted features, suitable weighting schemes are utilized to balance the different features. As each music clip may have rather complex emotions, a supervised multiclass label model is adopted to annotate emotions with emotion multinomial. Preliminary experimental results indicate that the proposed emotion model and retrieval approach is able to deliver good retrieval performance.

[1]  Changsheng Xu,et al.  Advances in Multimedia Information Processing - PCM 2008, 9th Pacific Rim Conference on Multimedia, Tainan, Taiwan, December 9-13, 2008. Proceedings , 2008, PCM.

[2]  Nicola Orio,et al.  Music Retrieval: A Tutorial and Review , 2006, Found. Trends Inf. Retr..

[3]  D. Västfjäll,et al.  Emotional responses to music: the need to consider underlying mechanisms. , 2008, The Behavioral and brain sciences.

[4]  Yi-Hsuan Yang,et al.  Music Emotion Classification: A Regression Approach , 2007, 2007 IEEE International Conference on Multimedia and Expo.

[5]  Gert R. G. Lanckriet,et al.  Modeling music and words using a multi-class naïve Bayes approach , 2006, ISMIR.

[6]  Petri Toiviainen,et al.  Prediction of Multidimensional Emotional Ratings in Music from Audio Using Multivariate Regression Models , 2009, ISMIR.

[7]  Gert R. G. Lanckriet,et al.  Semantic Annotation and Retrieval of Music and Sound Effects , 2008, IEEE Transactions on Audio, Speech, and Language Processing.

[8]  Sanghoon Jun,et al.  A fuzzy inference-based music emotion recognition system , 2008 .

[9]  Yi-Hsuan Yang,et al.  Toward Multi-modal Music Emotion Classification , 2008, PCM.

[10]  Yi-Hsuan Yang,et al.  Music emotion classification: a fuzzy approach , 2006, MM '06.

[11]  Lie Lu,et al.  Automatic mood detection and tracking of music audio signals , 2006, IEEE Transactions on Audio, Speech, and Language Processing.

[12]  Gert R. G. Lanckriet,et al.  Towards musical query-by-semantic-description using the CAL500 data set , 2007, SIGIR.