Multimedia information retrieval: music and audio

Music is an omnipresent topic in our daily lives, as almost everyone enjoys listening to his or her favorite tunes. Music information retrieval (MIR) is a research field that aims – among other things – at automatically extracting semantically meaningful information from various representations of music entities, such as a digital audio file, a band’s web page, a song’s lyrics, or a tweet about a microblogger’s current listening activity. A key approach in MIR is to describe music via computational features, which can be categorized into: music content, music context, and user context. The music content refers to features extracted from the audio signal, while information about musical entities not encoded in the signal (e.g., image of an artist or political background of a song) are referred to as music context. The user context, in contrast, includes environmental aspects as well as physical and mental activities of the music listener. MIR research has been seeing a paradigm shift over the last couple of years, as an increasing number of recent approaches and commercial technologies combine content-based techniques (focusing on the audio signal) with multimedia context data mined, e.g. from web sources and with user context information.

[1]  Emilia Gómez,et al.  Semantic audio content-based music recommendation and visualization based on user preference examples , 2013, Inf. Process. Manag..

[2]  Peter Knees,et al.  A music information system automatically generated via Web content mining techniques , 2011, Inf. Process. Manag..

[3]  Hiromasa Fujihara,et al.  Songle: A Web Service for Active Music Listening Improved by User Contributions , 2011, ISMIR.

[4]  Markus Schedl,et al.  Exploring Geospatial Music Listening Patterns in Microblog Data , 2012, Adaptive Multimedia Retrieval.

[5]  Peter Knees,et al.  The CoMIRVA Toolkit for Visualizing Music-Related Data , 2007, EuroVis.

[6]  Peter Knees,et al.  Exploring Music Collections in Virtual Landscapes , 2007, IEEE MultiMedia.

[7]  Emilia Gómez,et al.  Melody Extraction From Polyphonic Music Signals Using Pitch Contour Characteristics , 2012, IEEE Transactions on Audio, Speech, and Language Processing.

[8]  Peter Knees,et al.  Large-scale music exploration in hierarchically organized landscapes using prototypicality information , 2011, ICMR '11.

[9]  Masataka Goto,et al.  A chorus section detection method for musical audio signals and its application to a music listening station , 2006, IEEE Transactions on Audio, Speech, and Language Processing.

[10]  Perfecto Herrera,et al.  Mood Cloud : A Real-Time Music Mood Visualization Tool , 2008 .

[11]  Emilia Gómez Gutiérrez,et al.  Tonal description of music audio signals , 2006 .

[12]  Masataka Goto,et al.  Musicream: Integrated Music-Listening Interface for Active, Flexible, and Unexpected Encounters with Musical Pieces , 2009, J. Inf. Process..

[13]  Masataka Goto,et al.  Dancereproducer: An automatic mashup music video generation system by reusing dance video clips on the web , 2011 .

[14]  Kazuhito Yokoi,et al.  VocaListener and VocaWatcher: Imitating a human singer by using signal processing , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[15]  Peter Knees,et al.  Interactive Poster: Using CoMIRVA for Visualizing Similarities Between Music Artists , 2005, IEEE Visualization.