Case Study "Beatles Songs" - What can be Learned from Unreliable Music Alignments?

As a result of massive digitization efforts and the world wide web, there is an exploding amount of available digital data describing and representing music at various semantic levels and in diverse formats. For example, in the case of the Beatles songs, there are numerous recordings including an increasing number of cover songs and arrangements as well as MIDI data and other symbolic music representations. The general goal of music synchronization is to align the multiple information sources related to a given piece of music. This becomes a difficult problem when the various representations reveal significant differences in structure and polyphony, while exhibiting various types of artifacts. In this paper, we address the issue of how music synchronization techniques are useful for automatically revealing critical passages with significant difference between the two versions to be aligned. Using the corpus of the Beatles songs as test bed, we analyze the kind of differences occurring in audio and MIDI versions available for the songs.

[1]  Biing-Hwang Juang,et al.  Fundamentals of speech recognition , 1993, Prentice Hall signal processing series.

[2]  Geraint A. Wiggins,et al.  Treating Inherent Ambiguity in Ground Truth Data: Evaluating a Chord Labelling Algorithm , 2007 .

[3]  Meinard Müller,et al.  Towards an Efficient Algorithm for Automatic Score-to-Audio Synchronization , 2004, ISMIR.

[4]  Mark B. Sandler,et al.  Symbolic Representation of Musical Chords: A Proposed Syntax for Text Annotations , 2005, ISMIR.

[5]  George Tzanetakis,et al.  Polyphonic audio matching and alignment for music retrieval , 2003, 2003 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (IEEE Cat. No.03TH8684).

[6]  Christopher Raphael,et al.  A Hybrid Graphical Model for Aligning Polyphonic Audio with Musical Scores , 2004, ISMIR.

[7]  Ning Hu,et al.  Polyphonic Audio Matching for Score Following and Intelligent Audio Editors , 2003, ICMC.

[8]  Xavier Rodet,et al.  Improving polyphonic and poly-instrumental music to score alignment , 2003, ISMIR.

[9]  Meinard Müller,et al.  Automatic synchronization of music data in score-, MIDI- and PCM-format , 2003, ISMIR.

[10]  Meinard Müller,et al.  Information retrieval for music and motion , 2007 .

[11]  Meinard Müller,et al.  Joint Structure Analysis with Applications to Music Annotation and Synchronization , 2008, ISMIR.

[12]  Gregory H. Wakefield,et al.  Audio thumbnailing of popular music using chroma-based representations , 2005, IEEE Transactions on Multimedia.

[13]  Geraint A. Wiggins,et al.  High-level feature descriptors and corpus-based musicology: Techniques for modelling music cognition , 2008 .

[14]  Meinard Müller,et al.  An Efficient Multiscale Approach to Audio Synchronization , 2006, ISMIR.

[15]  Daniel Müllensiefen,et al.  Bayesian Model Selection for Harmonic Labelling , 2007 .

[16]  Xavier Serra,et al.  Chroma Binary Similarity and Local Alignment Applied to Cover Song Identification , 2008, IEEE Transactions on Audio, Speech, and Language Processing.

[17]  Meinard Müller,et al.  Path-constrained partial music synchronization , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.

[18]  Meinard Müller,et al.  Enhancing Similarity Matrices for Music Audio Analysis , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.

[19]  Gerhard Widmer,et al.  MATCH: A Music Alignment Tool Chest , 2005, ISMIR.

[20]  Sean R. Eddy,et al.  Biological Sequence Analysis: Probabilistic Models of Proteins and Nucleic Acids , 1998 .