Population Codes Representing Musical Timbre for High-Level fMRI Categorization of Music Genres

We present experimental evidence in support of distributed neural codes for timbre that are implicated in discrimination of musical styles. We used functional magnetic resonance imaging (fMRI) in humans and multivariate pattern analysis (MVPA) to identify activation patterns that encode the perception of rich music audio stimuli from five different musical styles. We show that musical styles can be automatically classified from population codes in bilateral superior temporal sulcus (STS). To investigate the possible link between the acoustic features of the auditory stimuli and neural population codes in STS, we conducted a representational similarity analysis and a multivariate regression-retrieval task. We found that the similarity structure of timbral features of our stimuli resembled the similarity structure of the STS more than any other type of acoustic feature. We also found that a regression model trained on timbral features outperformed models trained on other types of audio features. Our results show that human brain responses to complex, natural music can be differentiated by timbral audio features, emphasizing the importance of timbre in auditory perception.

[1]  Beth Logan,et al.  Mel Frequency Cepstral Coefficients for Music Modeling , 2000, ISMIR.

[2]  E. Formisano,et al.  Auditory Cortex Encodes the Perceptual Interpretation of Ambiguous Sound , 2011, The Journal of Neuroscience.

[3]  Tom Michael Mitchell,et al.  Predicting Human Brain Activity Associated with the Meanings of Nouns , 2008, Science.

[4]  Gary H. Glover,et al.  Neural Correlates of Timbre Change in Harmonic Sounds , 2002, NeuroImage.

[5]  Meinard Müller,et al.  Making chroma features more robust to timbre changes , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.

[6]  A. R. Jennings,et al.  Analysis of the spectral envelope of sounds by the human brain , 2005, NeuroImage.

[7]  George Tzanetakis,et al.  Musical genre classification of audio signals , 2002, IEEE Trans. Speech Audio Process..

[8]  Robert J. Zatorre,et al.  Contribution of the right temporal lobe to musical timbre discrimination , 1994, Neuropsychologia.

[9]  Sean M. Polyn,et al.  Beyond mind-reading: multi-voxel pattern analysis of fMRI data , 2006, Trends in Cognitive Sciences.

[10]  Keiji Tanaka,et al.  Matching Categorical Object Representations in Inferior Temporal Cortex of Man and Monkey , 2008, Neuron.

[11]  D Yves von Cramon,et al.  Distinct fMRI responses to laughter, speech, and sounds along the human peri-sylvian cortex. , 2005, Brain research. Cognitive brain research.

[12]  R. Zatorre,et al.  Melodic and harmonic discrimination following unilateral cerebral excision , 1988, Brain and Cognition.

[13]  Stephen José Hanson,et al.  Combinatorial codes in ventral temporal lobe for object recognition: Haxby (2001) revisited: is there a “face” area? , 2004, NeuroImage.

[14]  Richard Granger,et al.  Investigation of melodic contour processing in the brain using multivariate pattern-based fMRI , 2011, NeuroImage.

[15]  R W Cox,et al.  AFNI: software for analysis and visualization of functional magnetic resonance neuroimages. , 1996, Computers and biomedical research, an international journal.

[16]  Noël Staeren,et al.  Sound Categories Are Represented as Distributed Patterns in the Human Auditory Cortex , 2009, Current Biology.

[17]  Geoffrey M. Boynton,et al.  Efficient Design of Event-Related fMRI Experiments Using M-Sequences , 2002, NeuroImage.

[18]  Judith C. Brown,et al.  An efficient algorithm for the calculation of a constant Q transform , 1992 .

[19]  Kerry M. M. Walker,et al.  Sensitivity and Selectivity of Neurons in Auditory Cortex to the Pitch, Timbre, and Location of Sounds , 2010, The Neuroscientist : a review journal bringing neurobiology, neurology and psychiatry.