Distinct sensitivity to spectrotemporal modulation supports brain asymmetry for speech and melody

Speech versus music in the brain To what extent does the perception of speech and music depend on different mechanisms in the human brain? What is the anatomical basis underlying this specialization? Albouy et al. created a corpus of a cappella songs that contain both speech (semantic) and music (melodic) information and degraded each stimulus selectively in either the temporal or spectral domain. Degradation of temporal information impaired speech recognition but not melody recognition, whereas degradation of spectral information impaired melody recognition but not speech recognition. Brain scanning revealed a right-left asymmetry for speech and music. Classification of speech content occurred exclusively in the left auditory cortex, whereas classification of melodic content occurred only in the right auditory cortex. Science, this issue p. 1043 A human brain imaging study reveals that low-level acoustical cues support hemispheric lateralization of speech and music perception. Does brain asymmetry for speech and music emerge from acoustical cues or from domain-specific neural networks? We selectively filtered temporal or spectral modulations in sung speech stimuli for which verbal and melodic content was crossed and balanced. Perception of speech decreased only with degradation of temporal information, whereas perception of melodies decreased only with spectral degradation. Functional magnetic resonance imaging data showed that the neural decoding of speech and melodies depends on activity patterns in left and right auditory regions, respectively. This asymmetry is supported by specific sensitivity to spectrotemporal modulation rates within each region. Finally, the effects of degradation on perception were paralleled by their effects on neural classification. Our results suggest a match between acoustical properties of communicative signals and neural specializations adapted to that purpose.

[1]  Guillaume A. Rousselet,et al.  A statistical framework for neuroimaging data analysis based on mutual information estimated via a gaussian copula , 2016, bioRxiv.

[2]  David Poeppel,et al.  The analysis of speech in different temporal integration windows: cerebral lateralization as 'asymmetric sampling in time' , 2003, Speech Commun..

[3]  R. Zatorre,et al.  Structure and function of auditory cortex: music and speech , 2002, Trends in Cognitive Sciences.

[4]  M. Schönwiesner,et al.  Spectro-temporal modulation transfer function of single voxels in the human auditory cortex measured with high-resolution fMRI , 2009, Proceedings of the National Academy of Sciences.

[5]  Aniruddh D. Patel,et al.  Temporal modulations in speech and music , 2017, Neuroscience & Biobehavioral Reviews.

[6]  Josh H. McDermott,et al.  Distinct Cortical Pathways for Music and Speech Revealed by Hypothesis-Free Voxel Decomposition , 2015, Neuron.

[7]  Mireille Besson,et al.  Similar cerebral networks in language, music and song perception , 2010, NeuroImage.

[8]  John-Dylan Haynes,et al.  Valid population inference for information-based imaging: From the second-level t-test to prevalence inference , 2015, NeuroImage.

[9]  Frédéric E. Theunissen,et al.  The Modulation Transfer Function for Speech Intelligibility , 2009, PLoS Comput. Biol..

[10]  Powen Ru,et al.  Multiresolution spectrotemporal analysis of complex sounds. , 2005, The Journal of the Acoustical Society of America.

[11]  Mark A Pitt,et al.  How does context play a part in splitting words apart? Production and perception of word boundaries in casual speech. , 2012, Journal of memory and language.

[12]  J. Fritz,et al.  Rapid task-related plasticity of spectrotemporal receptive fields in primary auditory cortex , 2003, Nature Neuroscience.

[13]  David Poeppel,et al.  The cortical analysis of speech-specific temporal structure revealed by responses to sound quilts , 2015, Nature Neuroscience.

[14]  Jesper Andersson,et al.  A multi-modal parcellation of human cerebral cortex , 2016, Nature.

[15]  R V Shannon,et al.  Speech Recognition with Primarily Temporal Cues , 1995, Science.

[16]  M. Coltheart,et al.  Modularity of music processing , 2003, Nature Neuroscience.

[17]  Mark A. Chevillet,et al.  Functional Correlates of the Anterolateral Processing Hierarchy in Human Auditory Cortex , 2011, The Journal of Neuroscience.

[18]  Judit Gervain,et al.  Efficient Neural Coding in Auditory and Speech Perception , 2019, Trends in Neurosciences.

[19]  David Poeppel,et al.  Human Screams Occupy a Privileged Niche in the Communication Soundscape , 2015, Current Biology.

[20]  Sophie K. Scott,et al.  Cortical asymmetries in speech perception: what's wrong, what's right and what's left? , 2012, Trends in Cognitive Sciences.

[21]  David Poeppel,et al.  Spectrotemporal modulation provides a unifying framework for auditory cortical asymmetries , 2019, Nature Human Behaviour.

[22]  A. Boemio,et al.  Hierarchical and asymmetric temporal sensitivity in human auditory cortices , 2005, Nature Neuroscience.

[23]  Richard S. J. Frackowiak,et al.  Neurophysiological origin of human brain asymmetry for speech and language , 2010, Proceedings of the National Academy of Sciences.

[24]  M. Escabí,et al.  Spectral and temporal modulation tradeoff in the inferior colliculus. , 2010, Journal of neurophysiology.

[25]  J. Rauschecker,et al.  Cortical Representation of Natural Complex Sounds: Effects of Acoustic Features and Auditory Object Category , 2010, The Journal of Neuroscience.

[26]  Martin N. Hebart,et al.  The Decoding Toolbox (TDT): a versatile software package for multivariate analyses of functional imaging data , 2015, Front. Neuroinform..

[27]  M. Schönwiesner,et al.  Hemispheric asymmetry for spectral and temporal processing in the human antero‐lateral auditory belt cortex , 2005, The European journal of neuroscience.

[28]  R. Zatorre,et al.  Spectral and temporal processing in human auditory cortex. , 2001, Cerebral cortex.

[29]  A. Friederici Hierarchy processing in human neurobiology: how specific is it? , 2019, Philosophical Transactions of the Royal Society B.

[30]  Essa Yacoub,et al.  Encoding of Natural Sounds at Multiple Spectral and Temporal Resolutions in the Human Auditory Cortex , 2014, PLoS Comput. Biol..

[31]  Jonas Obleser,et al.  Bilateral Speech Comprehension Reflects Differential Sensitivity to Spectral and Temporal Features , 2008, The Journal of Neuroscience.