MIRAI: Multi-hierarchical, FS-Tree Based Music Information Retrieval System

With the fast booming of online music repositories, there is a need for content-based automatic indexing which will help users to find their favorite music objects in real time. Recently, numerous successful approaches on musical data feature extraction and selection have been proposed for instrument recognition in monophonic sounds. Unfortunately, none of these methods can be successfully applied to polyphonic sounds. Identification of music instruments in polyphonic sounds is still difficult and challenging, especially when harmonic partials are overlapping with each other. This has stimulated the research on music sound separation and new features development for content-based automatic music information retrieval. Our goal is to build a cooperative query answering system (QAS), for a musical database, retrieving from it all objects satisfying queries like "find all musical pieces in pentatonic scale with a viola and piano where viola is playing for minimum 20 seconds and piano for minimum 10 seconds". We use the database of musical sounds, containing almost 4000 sounds taken from the MUMs (McGill University Master Samples), as a vehicle to construct several classifiers for automatic instrument recognition. Classifiers showing the best performance are adopted for automatic indexing of musical pieces by instruments. Our musical database has an FS-tree (Frame Segment Tree) structure representation. The cooperativeness of QAS is driven by several hierarchical structures used for classifying musical instruments.

[1]  Zbigniew W. Ras,et al.  Maximum Likelihood Study for Sound Pattern Separation and Recognition , 2007, 2007 International Conference on Multimedia and Ubiquitous Engineering (MUE'07).

[2]  Pierre Comon,et al.  Independent component analysis, a survey of some algebraic methods , 1996, 1996 IEEE International Symposium on Circuits and Systems. Circuits and Systems Connecting the World. ISCAS 96.

[3]  Parke Godfrey,et al.  Minimization in Cooperative Response to Failing Database Queries , 1994, Int. J. Cooperative Inf. Syst..

[4]  J. Flanagan Speech Analysis, Synthesis and Perception , 1971 .

[5]  Zbigniew W. Ras,et al.  Analysis of Sound Features for Music Timbre Recognition , 2007, 2007 International Conference on Multimedia and Ubiquitous Engineering (MUE'07).

[6]  Ronald R. Coifman,et al.  Aspects of Pitch-Tracking and Timbre Separation: Feature Detection in Digital Audio Using Adapted Local Trigonometric Bases and Wavelet Packages , 1995, ICMC.

[7]  Zhang Peng,et al.  Research on the Integration in E-government Based on Multi-agent , 2006, 2006 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology Workshops.

[8]  Zbigniew W. Ras,et al.  Blind Signal Separation of Similar Pitches and Instruments in a Noisy Polyphonic Domain , 2006, ISMIS.

[9]  Ronald W. Schafer,et al.  Digital Processing of Speech Signals , 1978 .

[10]  Jonathan Berger,et al.  SONART : THE SONIFICATION APPLICATION RESEARCH TOOLBOX , 2002 .

[11]  B. S. Manjunath,et al.  Introduction to mpeg-7 , 2002 .

[12]  Shusaku Tsumoto,et al.  Foundations of Intelligent Systems, 15th International Symposium, ISMIS 2005, Saratoga Springs, NY, USA, May 25-28, 2005, Proceedings , 2005, ISMIS.

[13]  Youngmoo E. Kim,et al.  Musical instrument identification: A pattern‐recognition approach , 1998 .

[14]  Xavier Serra,et al.  Towards Instrument Segmentation for Music Content Description: a Critical Review of Instrument Classification Techniques , 2000, ISMIR.

[15]  Andrzej Czyzewski,et al.  Representing Musical Instrument Sounds for Their Automatic Classification , 2001 .

[16]  Terry Gaasterland Cooperative Answering through Controlled Query Relaxation , 1997, IEEE Expert.

[17]  Alicja Wieczorkowska Musical Sound Classification based on Wavelet Analysis , 2001, Fundam. Informaticae.

[18]  Alexander H. Waibel,et al.  Towards Unrestricted Lip Reading , 2000, Int. J. Pattern Recognit. Artif. Intell..

[19]  B. S. Manjunath,et al.  Introduction to MPEG-7: Multimedia Content Description Interface , 2002 .

[20]  Edward A. Lee,et al.  Adaptive Signal Models: Theory, Algorithms, and Audio Applications , 1998 .

[21]  Ichiro Fujinaga,et al.  Realtime Recognition of Orchestral Instruments , 2000, International Conference on Mathematics and Computing.

[22]  J C Brown,et al.  Feature dependence in the automatic identification of musical woodwind instruments. , 2001, The Journal of the Acoustical Society of America.

[23]  Albert S. Bregman,et al.  The Auditory Scene. (Book Reviews: Auditory Scene Analysis. The Perceptual Organization of Sound.) , 1990 .

[24]  Zbigniew W. Ras,et al.  Multi-way Hierarchic Classification of Musical Instrument Sounds , 2007, 2007 International Conference on Multimedia and Ubiquitous Engineering (MUE'07).

[25]  Zbigniew W. Ras,et al.  Differentiated harmonic feature analysis on music information retrieval for instrument recognition , 2006, 2006 IEEE International Conference on Granular Computing.