Classification of musical styles using liquid state machines

Music Information Retrieval (MIR) is an interdisciplinary field that facilitates indexing and content-based organization of music databases. Music classification and clustering is one of the major topics in MIR. Music can be defined as ‘organized sound’. The highly ordered temporal structure of music suggests it should be amendable to analysis by a novel spiking neural network paradigm: the liquid state machine (LSM). Unlike conventional statistical approaches that require the presence of static input data, the LSM has a unique ability to classify music in real-time, due to its dynamics and fading-memory. This paper investigates the performance of an LSM in classifying musical styles (ragtime vs. classical), as well as its ability to distinguish music from note sequences without temporal structure. The results show that the LSM performs admirably in this task.

[1]  David S. Watson,et al.  A Machine Learning Approach to Musical Style Recognition , 1997, ICMC.

[2]  Marco Wiering,et al.  Democratic Liquid State Machines for Music Recognition , 2008, Speech, Audio, Image and Biomedical Signal Processing using Neural Networks.

[3]  Henry Markram,et al.  Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations , 2002, Neural Computation.

[4]  Prashant Joshi,et al.  From memory-based decisions to decision-based movements: A model of interval discrimination followed by action selection , 2007, Neural Networks.

[5]  Pedro Ponce de Len,et al.  Pattern Recognition Approach for Music Style Identification Using Shallow Statistical Descriptors , 2007, IEEE Trans. Syst. Man Cybern. Part C.

[6]  Hojjat Adeli,et al.  Spiking Neural Networks , 2009, Int. J. Neural Syst..

[7]  Chris Eliasmith,et al.  Neural Engineering: Computation, Representation, and Dynamics in Neurobiological Systems , 2004, IEEE Transactions on Neural Networks.

[8]  P.P. de Leon,et al.  Pattern Recognition Approach for Music Style Identification Using Shallow Statistical Descriptors , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[9]  Henry Markram,et al.  Computer models and analysis tools for neural microcircuits , 2003 .

[10]  Jonathan Foote,et al.  An overview of audio information retrieval , 1999, Multimedia Systems.

[11]  Petri Toiviainen,et al.  MIDI toolbox : MATLAB tools for music research , 2004 .

[12]  Henry Markram,et al.  On the computational power of circuits of spiking neurons , 2004, J. Comput. Syst. Sci..

[13]  Belinda Thom,et al.  Unsupervised Learning and Interactive Jazz/Blues Improvisation , 2000, AAAI/IAAI.

[14]  H. Markram,et al.  Organizing principles for a diversity of GABAergic interneurons and synapses in the neocortex. , 2000, Science.

[15]  Marco Wiering,et al.  Speech, Audio, Image and Biomedical Signal Processing using Neural Networks , 2008 .