Machine recognition of timbre using steady-state tone of acoustic musical instruments

Recent experiments indicate that steady-state portion of an acoustic musical instrument may be sufficient for timbre recognition. Here a computer-based classifier was used to recognize very short samples of steady-state tones. Gregory Sandell’s SHARC database consisted of 39 different timbre (23 orchestral instruments, some with different articulations) played at different pitches (total of 1338 spectra) were used as the samples for an exemplar-based learning system that incorporates k-nearest neighbor classifier with genetic algorithm. The latter is used to find the optimal set of weights for the features to improve the classification. The features calculated from the spectral data of the steady-state portion of the instrumental sound included centroid and other moments, such as skewness and kurtosis. As expected the centroid alone was the best single feature with a recognition rate of 20%, which is much better than chance (2.5%). The best results were obtained using seven features: the fundamental, the integral of the spectrum, the centroid, the standard deviation, the skewness, and the first two harmonic partials. What was surprising was that the recognition varied greatly between instruments. While the French horn and the muted trumpet were recognized over 90%, the recognition of other instruments, such as the cello with martele (18%), and the violin pizzicato (14%) were very poor. The average overall was 50.3%.

[1]  Mehdi Hatamian,et al.  Optical character recognition by the method of moments , 1987 .

[2]  K. Berger Some Factors in the Recognition of Timbre , 1964 .

[3]  Paolo Prandoni,et al.  Sonological models for timbre characterization , 1997 .

[4]  R. Nosofsky Attention, similarity, and the identification-categorization relationship. , 1986, Journal of experimental psychology. General.

[5]  Shlomo Dubnov,et al.  Polyspectra as measures of sound texture and timbre , 1997 .

[6]  Roger A. Kendall,et al.  The role of acoustic signal partitions in listener categorization of musical phrases. , 1986 .

[7]  Douglas L. Medin,et al.  Context theory of classification learning. , 1978 .

[8]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[9]  J. F. Corso,et al.  Timbre Cues and the Identification of Musical Instruments , 1962 .

[10]  Ichiro Fujinaga Exemplar-based Learning in Adaptive Optical Music Recognition System , 1996, ICMC.

[11]  Shihab A. Shamma,et al.  Representation of musical timbre in the auditory cortex , 1997 .

[12]  Jack Sklansky,et al.  A note on genetic algorithms for large-scale feature selection , 1989, Pattern Recognition Letters.

[13]  Lawrence Davis,et al.  Hybridizing the Genetic Algorithm and the K Nearest Neighbors Classification Algorithm , 1991, ICGA.

[14]  Jan M. Van Campenhout,et al.  On the Possible Orderings in the Measurement Selection Problem , 1977, IEEE Transactions on Systems, Man, and Cybernetics.

[15]  J. Grey Timbre discrimination in musical patterns. , 1978, The Journal of the Acoustical Society of America.

[16]  R. Nosofsky American Psychological Association, Inc. Choice, Similarity, and the Context Theory of Classification , 2022 .

[17]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .