Methodological Considerations Concerning Manual Annotation Of Musical Audio In Function Of Algorithm Development

In research on musical audio-mining, annotated music databases are needed which allow the development of computational tools that extract from the musical audiostream the kind of high-level content that users can deal with in Music Information Retrieval (MIR) contexts. The notion of musical content, and therefore the notion of annotation, is ill-defined, however, both in the syntactic and semantic sense. As a consequence, annotation has been approached from a variety of perspectives (but mainly linguistic-symbolic oriented), and a general methodology is lacking. This paper is a step towards the definition of a general framework for manual annotation of musical audio in function of a computational approach to musical audio-mining that is based on algorithms that learn from annotated data.

[1]  Masataka Goto,et al.  RWC Music Database: Music genre database and musical instrument sound database , 2003, ISMIR.

[2]  George Tzanetakis,et al.  Experiments in computer-assisted annotation of audio , 2000 .

[3]  François Pachet,et al.  Evolving Automatically High-Level Music Descriptors from Acoustic Signals , 2003, CMMR.

[4]  Mark Liberman,et al.  A formal framework for linguistic annotation , 1999, Speech Commun..

[5]  Mark Notess,et al.  Timeliner: Building a Learning Tool into a Digital Music Library , 2004 .

[6]  Nicholas J. Belkin,et al.  Categories of Music Description and Search Terms and Phrases Used by Non-Music Experts , 2002, ISMIR.

[7]  Paul Boersma,et al.  Praat, a system for doing phonetics by computer , 2002 .

[8]  Marc Leman,et al.  The MAMI query-by-voice experiment: collecting and annotating vocal queries for music information retrieval , 2003, ISMIR.

[9]  Steffen Staab,et al.  On deep annotation , 2003, WWW '03.

[10]  William P. Birmingham,et al.  MUSART: Music Retrieval Via Aural Queries , 2001, ISMIR.

[11]  Gerhard Widmer,et al.  Exploring Music Collections by Browsing Different Views , 2004, Computer Music Journal.

[12]  Stephan Baumann,et al.  AN ECOLOGICAL APPROACH TO MULTIMODAL SUBJECTIVE MUSIC SIMILARITY PERCEPTION , 2004 .

[13]  Antonio Camurri,et al.  Gesture-Based Communication in Human-Computer Interaction , 2003, Lecture Notes in Computer Science.

[14]  C. Krumhansl,et al.  Measuring and Modeling Real-Time Responses to Music: The Dynamics of Tonality Induction , 2003, Perception.

[15]  Marc Leman,et al.  Correlation of Gestural Musical Audio Cues and Perceived Expressive Qualities , 2003, Gesture Workshop.

[16]  Lisa Harper,et al.  Emerging requirements for multi-modal annotation and analysis tools , 2001, INTERSPEECH.

[17]  Steffen Pauws,et al.  CubyHum: a fully operational "query by humming" system , 2002, ISMIR.

[18]  Marc Leman,et al.  An auditory model based transriber of vocal queries , 2003, ISMIR.

[19]  Brian C. Smith,et al.  Query By Humming , 2001 .

[20]  Ian H. Witten,et al.  The New Zealand Digital Library MELody inDEX , 1997, D Lib Mag..

[21]  M. Tomasello The Cultural Origins of Human Cognition , 2000 .