Application of Temporal Descriptors to Musical Instrument Sound Recognition

An automatic content extraction from multimedia files is recently being extensively explored. However, an automatic content description of musical sounds has not been broadly investigated and still needs an intensive research. In this paper, we investigate how to optimize sound representation in terms of musical instrument recognition purposes. We propose to trace trends in the evolution of values of MPEG-7 descriptors in time, as well as their combinations. Described process is a typical example of KDD application, consisting of data preparation, feature extraction and decision model construction. Discussion of efficiency of applied classifiers illustrates capabilities of possible progress in the optimization of sound representation. We believe that further research in this area would provide background for an automatic multimedia content description.

[1]  Marcin S. Szczuka,et al.  RSES and RSESlib - A Collection of Tools for Rough Set Computations , 2000, Rough Sets and Current Trends in Computing.

[2]  G. C. Tiao,et al.  Bayesian inference in statistical analysis , 1973 .

[3]  Jakub Wroblewski,et al.  Ensembles of Classifiers Based on Approximate Reducts , 2001, Fundam. Informaticae.

[4]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[5]  Shigeo Ando,et al.  Statistical study of spectral parameters in musical instrument tones , 1993 .

[6]  Stephen McAdams,et al.  Instrument Sound Description in the Context of MPEG-7 , 2000, ICMC.

[7]  Perry R. Cook,et al.  An automatic musical frequency detection and MIDI control system for brass instruments , 1992 .

[8]  Janusz Zalewski,et al.  Rough sets: Theoretical aspects of reasoning about data , 1996 .

[9]  日高 恒義,et al.  ISO/IEC JTC1 SC29/WG11 MPEG-2久里浜主観評価テスト報告 , 1992 .

[10]  Anssi Klapuri,et al.  Musical instrument recognition using cepstral coefficients and temporal features , 2000, 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.00CH37100).

[11]  Dominik Slezak,et al.  Classification Algorithms Based on Linear Combinations of Features , 1999, PKDD.

[12]  Ramakrishnan Srikant,et al.  Fast Algorithms for Mining Association Rules in Large Databases , 1994, VLDB.

[13]  Sinh Hoa Nguyen,et al.  Regularity analysis and its applications in data mining , 2000 .

[14]  Petri Toiviainen Optimizing Self-Organizing Timbre Maps: Two Approaches , 1996, Joint International Conference on Cognitive and Systematic Musicology.

[15]  Heikki Mannila,et al.  Fast Discovery of Association Rules , 1996, Advances in Knowledge Discovery and Data Mining.

[16]  David Cooper,et al.  A Monophonic Pitch-Tracking Algorithm Based on Waveform Periodicity Determinations Using Landmark Points , 1996 .

[17]  Xavier Serra,et al.  Towards Instrument Segmentation for Music Content Description: a Critical Review of Instrument Classification Techniques , 2000, ISMIR.

[18]  Heikki Mannila,et al.  Discovery of Frequent Episodes in Event Sequences , 1997, Data Mining and Knowledge Discovery.

[19]  Jürgen Herre,et al.  MPEG-7 and MPEG-7 audio: An overview , 2001 .

[20]  J C Brown Computer identification of musical instruments using pattern recognition with cepstral coefficients as features. , 1999, The Journal of the Acoustical Society of America.

[21]  Piotr Synak,et al.  Temporal Templates and Analysis of Time Related Data , 2000, Rough Sets and Current Trends in Computing.

[22]  J C Brown,et al.  Feature dependence in the automatic identification of musical woodwind instruments. , 2001, The Journal of the Acoustical Society of America.

[23]  Xavier Rodet,et al.  Estimation of fundamental frequency of musical sound signals , 1991, [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing.

[24]  Zbigniew W. Ras,et al.  Audio Content Description in Sound Databases , 2001, Web Intelligence.

[25]  Craig Stuart Sapp,et al.  Efficient Pitch Detection Techniques for Interactive Music , 2001, ICMC.

[26]  Andrzej Czyzewski,et al.  Representing Musical Instrument Sounds for Their Automatic Classification , 2001 .

[27]  James W. Beauchamp,et al.  Detection of Musical Pitch from Recorded Solo Performances , 1993 .

[28]  Marcin S. Szczuka,et al.  A New Version of Rough Set Exploration System , 2002, Rough Sets and Current Trends in Computing.

[29]  Alicja Wieczorkowska,et al.  Rough Sets as A Tool for Audio Signal Classification , 1999, ISMIS.

[30]  Jan G. Bazan,et al.  Rough set algorithms in classification problem , 2000 .

[31]  D Cooper,et al.  A Monophonic Pitch Tracking Algorithm , 1994 .

[32]  Piotr Synak,et al.  KDD-Based Approach to Musical Instrument Sound Recognition , 2002, ISMIS.

[33]  Lech Polkowski,et al.  Rough Sets in Knowledge Discovery 2 , 1998 .

[34]  Youngmoo E. Kim,et al.  Musical instrument identification: A pattern‐recognition approach , 1998 .

[35]  Ramakrishnan Srikant,et al.  Fast algorithms for mining association rules , 1998, VLDB 1998.

[36]  Hung Son Nguyen,et al.  Rough Set Data Analysis in the KDD Process , 2000 .

[37]  Pedro Cano,et al.  Automatic Segmentation for Music Classification using Competitive Hidden Markov Models , 2000, ISMIR.

[38]  Judith C. Brown,et al.  Musical frequency tracking using the methods of conventional and , 1991 .

[39]  Piero Cosi,et al.  Auditory modelling and self‐organizing neural networks for timbre classification , 1994 .

[40]  Hiroshi Motoda,et al.  Feature Extraction, Construction and Selection: A Data Mining Perspective , 1998 .

[41]  Ichiro Fujinaga,et al.  Realtime Recognition of Orchestral Instruments , 2000, International Conference on Mathematics and Computing.