Human Neuromagnetic Steady-State Responses to Amplitude-Modulated Tones, Speech, and Music

Objectives: Auditory steady-state responses that can be elicited by various periodic sounds inform about subcortical and early cortical auditory processing. Steady-state responses to amplitude-modulated pure tones have been used to scrutinize binaural interaction by frequency-tagging the two ears’ inputs at different frequencies. Unlike pure tones, speech and music are physically very complex, as they include many frequency components, pauses, and large temporal variations. To examine the utility of magnetoencephalographic (MEG) steady-state fields (SSFs) in the study of early cortical processing of complex natural sounds, the authors tested the extent to which amplitude-modulated speech and music can elicit reliable SSFs. Design: MEG responses were recorded to 90-s-long binaural tones, speech, and music, amplitude-modulated at 41.1 Hz at four different depths (25, 50, 75, and 100%). The subjects were 11 healthy, normal-hearing adults. MEG signals were averaged in phase with the modulation frequency, and the sources of the resulting SSFs were modeled by current dipoles. After the MEG recording, intelligibility of the speech, musical quality of the music stimuli, naturalness of music and speech stimuli, and the perceived deterioration caused by the modulation were evaluated on visual analog scales. Results: The perceived quality of the stimuli decreased as a function of increasing modulation depth, more strongly for music than speech; yet, all subjects considered the speech intelligible even at the 100% modulation. SSFs were the strongest to tones and the weakest to speech stimuli; the amplitudes increased with increasing modulation depth for all stimuli. SSFs to tones were reliably detectable at all modulation depths (in all subjects in the right hemisphere, in 9 subjects in the left hemisphere) and to music stimuli at 50 to 100% depths, whereas speech usually elicited clear SSFs only at 100% depth. The hemispheric balance of SSFs was toward the right hemisphere for tones and speech, whereas SSFs to music showed no lateralization. In addition, the right lateralization of SSFs to the speech stimuli decreased with decreasing modulation depth. Conclusions: The results showed that SSFs can be reliably measured to amplitude-modulated natural sounds, with slightly different hemispheric lateralization for different carrier sounds. With speech stimuli, modulation at 100% depth is required, whereas for music the 75% or even 50% modulation depths provide a reasonable compromise between the signal-to-noise ratio of SSFs and sound quality or perceptual requirements. SSF recordings thus seem feasible for assessing the early cortical processing of natural sounds.

[1]  O Bertrand,et al.  Analysis of speech sounds is left-hemisphere predominant at 100-150ms after sound onset. , 1999, Neuroreport.

[2]  Matthias M. Müller,et al.  Sustained selective intermodal attention modulates processing of language-like stimuli , 2011, Experimental Brain Research.

[3]  A. Friederici,et al.  Lateralization of auditory language functions: A dynamic dual pathway model , 2004, Brain and Language.

[4]  B. Ross A novel type of auditory responses: temporal dynamics of 40-Hz steady-state responses induced by changes in sound localization. , 2008, Journal of neurophysiology.

[5]  A. Lindell In Your Right Mind: Right Hemisphere Contributions to Language Processing and Production , 2006, Neuropsychology Review.

[6]  A. Rees,et al.  Steady-state evoked responses to sinusoidally amplitude-modulated sounds recorded in man , 1986, Hearing Research.

[7]  S. Makeig,et al.  A 40-Hz auditory potential recorded from the human scalp. , 1981, Proceedings of the National Academy of Sciences of the United States of America.

[8]  Robert J Zatorre,et al.  Neural specializations for speech and pitch: moving beyond the dichotomies , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.

[9]  Terence W Picton,et al.  Auditory Steady-State Responses and Word Recognition Scores in Normal-Hearing and Hearing-Impaired Adults , 2004, Ear and hearing.

[10]  Jan Wouters,et al.  Auditory Steady State Cortical Responses Indicate Deviant Phonemic-Rate Processing in Adults With Dyslexia , 2012, Ear and hearing.

[11]  R. Hari,et al.  Binaural interaction and the octave illusion. , 2012, The Journal of the Acoustical Society of America.

[12]  M Hämäläinen,et al.  Neuromagnetic steady-state responses to auditory stimuli. , 1989, The Journal of the Acoustical Society of America.

[13]  A. Friederici The brain basis of language processing: from structure to function. , 2011, Physiological reviews.

[14]  C. Pantev,et al.  Tonotopic organization of the sources of human auditory steady-state responses , 1996, Hearing Research.

[15]  Shigeyuki Kuwada,et al.  Scalp potentials of normal and hearing-impaired subjects in response to sinusoidally amplitude-modulated tones , 1986, Hearing Research.

[16]  Riitta Hari,et al.  Human ROBO1 Regulates Interaural Interaction in Auditory Pathways , 2012, The Journal of Neuroscience.

[17]  T. Picton,et al.  Human auditory steady-state responses: Respuestas auditivas de estado estable en humanos , 2003, International journal of audiology.

[18]  R. Zatorre,et al.  Spectral and temporal processing in human auditory cortex. , 2001, Cerebral cortex.

[19]  M. Jung-Beeman Bilateral brain processes for comprehending natural language , 2005, Trends in Cognitive Sciences.

[20]  R. Zatorre,et al.  Structure and function of auditory cortex: music and speech , 2002, Trends in Cognitive Sciences.

[21]  R V Shannon,et al.  Speech Recognition with Primarily Temporal Cues , 1995, Science.

[22]  R. Ilmoniemi,et al.  Magnetoencephalography-theory, instrumentation, and applications to noninvasive studies of the working human brain , 1993 .

[23]  M. Scherg,et al.  Deconvolution of 40 Hz steady-state fields reveals two overlapping source activities of the human auditory cortex , 1999, Clinical Neurophysiology.

[24]  Terence W Picton,et al.  Recording Auditory Steady-State Responses in Young Infants , 2004, Ear and Hearing.

[25]  M. Tervaniemi,et al.  Lateralization of auditory-cortex functions , 2003, Brain Research Reviews.

[26]  C Pantev,et al.  A high-precision magnetoencephalographic study of human auditory steady-state responses to amplitude-modulated tones. , 2000, The Journal of the Acoustical Society of America.

[27]  Tapani Ristaniemi,et al.  From Vivaldi to Beatles and back: Predicting lateralized brain responses to music , 2013, NeuroImage.

[28]  Riitta Hari,et al.  Binaural interaction in the human auditory cortex revealed by neuromagnetic frequency tagging: no effect of stimulus intensity , 2003, Hearing Research.

[29]  C Pantev,et al.  Right hemispheric laterality of human 40 Hz auditory steady-state responses. , 2005, Cerebral cortex.

[30]  R. C. Oldfield The assessment and analysis of handedness: the Edinburgh inventory. , 1971, Neuropsychologia.

[31]  Riitta Hari,et al.  Neuromagnetic Responses to Frequency-Tagged Sounds: A New Method to Follow Inputs from Each Ear to the Human Auditory Cortex during Binaural Hearing , 2002, The Journal of Neuroscience.

[32]  T W Picton,et al.  Potentials evoked by the sinusoidal modulation of the amplitude or frequency of a tone. , 1987, The Journal of the Acoustical Society of America.

[33]  R Hari,et al.  Evidence for cortical origin of the 40 Hz auditory evoked response in man. , 1987, Electroencephalography and clinical neurophysiology.

[34]  R. C. Oldfield THE ASSESSMENT AND ANALYSIS OF HANDEDNESS , 1971 .

[35]  T. Griffiths The Neural Processing of Complex Sounds , 2001, Annals of the New York Academy of Sciences.