Spectral and Temporal Periodicity Representations of Rhythm for the Automatic Classification of Music Audio Signal

In this paper, we study the spectral and temporal periodicity representations that can be used to describe the characteristics of the rhythm of a music audio signal. A continuous-valued energy-function representing the onset positions over time is first extracted from the audio signal. From this function we compute at each time a vector which represents the characteristics of the local rhythm. Four feature sets are studied for this vector. They are derived from the amplitude of the discrete Fourier transform (DFT), the auto-correlation function (ACF), the product of the DFT and the ACF interpolated on a hybrid lag/frequency axis and the concatenated DFT and ACF coefficients. Then the vectors are sampled at some specific frequencies, which represent various ratios of the local tempo. The ability of these periodicity representations to describe the rhythm characteristics of an audio item is evaluated through a classification task. In this, we test the use of the periodicity representations alone, combined with tempo information and combined with a proposed set of rhythm features. The evaluation is performed using annotated and estimated tempo. We show that using such simple periodicity representations allows achieving high recognition rates at least comparable to previously published results.

[1]  Jarno Seppänen,et al.  Joint Beat & Tatum Tracking from Music Signals , 2006, ISMIR.

[2]  Daniel Gärtner,et al.  Improving Rhythmic Similarity Computation by Beat Histogram Transformations , 2009, ISMIR.

[3]  Sergios Theodoridis,et al.  Music Retrieval by Rhythmic Similarity Applied on Greek and African Traditional Music , 2007, ISMIR.

[4]  Anssi Klapuri,et al.  Measuring the similarity of Rhythmic Patterns , 2002, ISMIR.

[5]  Geoffroy Peeters,et al.  Template-Based Estimation of Time-Varying Tempo , 2007, EURASIP J. Adv. Signal Process..

[6]  Mark A. Hall,et al.  Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning , 1999, ICML.

[7]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[8]  Gerhard Widmer,et al.  Exploring Music Collections by Browsing Different Views , 2004, Computer Music Journal.

[9]  Geoffroy Peeters,et al.  Music Pitch Representation by Periodicity Measures Based on Combined Temporal and Spectral Representations , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.

[10]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[11]  Jonathan Foote,et al.  Audio Retrieval by Rhythmic Similarity , 2002, ISMIR.

[12]  Xavier Rodet,et al.  HIERARCHICAL GAUSSIAN TREE WITH INERTIA RATIO MAXIMIZATION FOR THE CLASSIFICATION OF LARGE MUSICAL INSTRUMENT DATABASES , 2003 .

[13]  G. Peeters,et al.  Non-Stationary Analysis/Synthesis using Spectrum Peak Shape Distortion, Phase and Reassignement , 1999 .

[14]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[15]  Gerhard Widmer,et al.  Evaluating Rhythmic descriptors for Musical Genre Classification , 2004 .

[16]  Yong Wang,et al.  Using Model Trees for Classification , 1998, Machine Learning.

[17]  George Tzanetakis,et al.  Analyzing Afro-Cuban Rhythms using Rotation-Aware Clave Template Matching with Dynamic Programming , 2008, ISMIR.

[18]  Søren Holdt Jensen,et al.  A tempo-insensitive representation of rhythmic patterns , 2009, 2009 17th European Signal Processing Conference.

[19]  Yannis Stylianou,et al.  Rhythmic similarity of music based on dynamic periodicity warping , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.

[20]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[21]  Christian Dittmar,et al.  Drum Pattern Based Genre Classification of Popular Music , 2004 .

[22]  Simon Dixon,et al.  Dance music classification: A tempo-based approach , 2004, ISMIR.

[23]  Shingo Uchihashi,et al.  The beat spectrum: a new approach to rhythm analysis , 2001, IEEE International Conference on Multimedia and Expo, 2001. ICME 2001..

[24]  Ian H. Witten,et al.  Generating Accurate Rule Sets Without Global Optimization , 1998, ICML.

[25]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[26]  Hirokazu Kameoka,et al.  A Real-time Equalizer of Harmonic and Percussive Components in Music Signals , 2008, ISMIR.

[27]  Gerhard Widmer,et al.  Towards Characterisation of Music via Rhythmic Patterns , 2004, ISMIR.

[28]  George Tzanetakis,et al.  Audio genre classification using percussive pattern clustering combined with timbral features , 2009, 2009 IEEE International Conference on Multimedia and Expo.

[29]  M. Puckette,et al.  Calculation of a "narrowed" autocorrelation function. , 1989, The Journal of the Acoustical Society of America.

[30]  Yannis Stylianou,et al.  A scale transform based method for rhythmic similarity of music , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.

[31]  George Tzanetakis,et al.  Musical genre classification of audio signals , 2002, IEEE Trans. Speech Audio Process..