Correcting MEG Artifacts Caused by Overt Speech

Recording brain activity during speech production using magnetoencephalography (MEG) can help us to understand the dynamics of speech production. However, these measurements are challenging due to the induced artifacts coming from several sources such as facial muscle activity, lower jaw and head movements. Here, we aimed to characterise speech-related artifacts and subsequently present an approach to remove these artifacts from MEG data. We recorded MEG from 11 healthy participants while they pronounced various syllables in different loudness. Head positions/orientations were extracted during speech production to investigate its role in MEG distortions. Finally, we present an artifact rejection approach using the combination of regression analysis and signal space projection (SSP) in order to correct the induced artifact from MEG data. Our results show that louder speech leads to stronger head movements and stronger MEG distortions. Our proposed artifact rejection approach could successfully remove the speech-related artifact and retrieve the underlying neurophysiological signals. As the presented artifact rejection approach was shown to remove artifacts induced by overt speech in the MEG, it will facilitate research addressing the neural basis of speech production with MEG.

[1]  Jayaram Chandrashekar,et al.  Sequential Processing of Lexical, Grammatical, and Phonological Information Within Broca's Area , 2009 .

[2]  R. Salmelin,et al.  Dynamic reconfiguration of the language network preceding onset of speech in picture naming , 2014, Human brain mapping.

[3]  Sabine Van Huffel,et al.  Removal of Muscle Artifacts from EEG Recordings of Spoken Language Production , 2010, Neuroinformatics.

[4]  F. Pulvermüller,et al.  Spatiotemporal Signatures of Large-Scale Synfire Chains for Speech Processing as Revealed by MEG , 2008, Cerebral cortex.

[5]  Mia Liljeström,et al.  Task- and stimulus-related cortical networks in language production: Exploring similarity of MEG- and fMRI-derived functional connectivity , 2015, NeuroImage.

[6]  Hideaki Tanaka,et al.  Error-related brain potentials elicited by vocal errors , 2001, Neuroreport.

[7]  Markus Butz,et al.  Rejecting deep brain stimulation artefacts from MEG data using ICA and mutual information , 2016, Journal of Neuroscience Methods.

[8]  Robin A. A. Ince,et al.  Representational interactions during audiovisual speech entrainment: Redundancy in left posterior superior temporal gyrus and synergy in left motor cortex , 2018, PLoS biology.

[9]  Robert Oostenveld,et al.  FieldTrip: Open Source Software for Advanced Analysis of MEG, EEG, and Invasive Electrophysiological Data , 2010, Comput. Intell. Neurosci..

[10]  G. Hickok Computational neuroanatomy of speech production , 2012, Nature Reviews Neuroscience.

[11]  F.-Xavier Alario,et al.  Intra-Cranial Recordings of Brain Activity During Language Production , 2011, Front. Psychology.

[12]  Jan Kujala,et al.  The right hemisphere is highlighted in connected natural speech production and perception , 2017, NeuroImage.

[13]  Asif A. Ghazanfar,et al.  The Natural Statistics of Audiovisual Speech , 2009, PLoS Comput. Biol..

[14]  D. Poeppel,et al.  Neural basis of speech perception. , 2015, Handbook of clinical neurology.

[15]  Richard J. Davidson,et al.  Electromyogenic Artifacts and Electroencephalographic Inferences , 2009, Brain Topography.

[16]  Greg Gibson,et al.  Rare and common variants: twenty arguments , 2012, Nature Reviews Genetics.

[17]  Gregor Thut,et al.  Lip movements entrain the observers’ low-frequency brain oscillations to facilitate speech intelligibility , 2016, eLife.

[18]  Olaf Hauk,et al.  Electroencephalographic activity over temporal brain areas during phonological encoding in picture naming , 2000, Clinical Neurophysiology.

[19]  R. Salmelin,et al.  Motor cortex dynamics in visuomotor production of speech and non-speech mouth movements. , 2006, Cerebral cortex.

[20]  R. Ilmoniemi,et al.  Signal-space projection method for separating MEG or EEG into components , 1997, Medical and Biological Engineering and Computing.

[21]  Joachim Gross,et al.  Good practice for conducting and reporting MEG research , 2013, NeuroImage.

[22]  E. Brown,et al.  Left-Lateralized Contributions of Saccades to Cortical Activity During a One-Back Word Recognition Task , 2017, bioRxiv.

[23]  David Poeppel,et al.  Chapter 25 – Neural Basis of Speech Perception , 2016 .

[24]  S. Muthukumaraswamy High-frequency brain activity and muscle artifacts in MEG/EEG: a review and recommendations , 2013, Front. Hum. Neurosci..

[25]  J. Gross Magnetoencephalography in Cognitive Neuroscience: A Primer , 2019, Neuron.

[26]  M. Merzenich,et al.  Modulation of the Auditory Cortex during Speech: An MEG Study , 2002, Journal of Cognitive Neuroscience.

[27]  Mia Liljeström,et al.  MEG evoked responses and rhythmic activity provide spatiotemporally complementary measures of neural activity in language production , 2012, NeuroImage.

[28]  M. Kutas,et al.  Electrophysiological estimates of the time course of semantic and phonological encoding during implicit picture naming. , 2000, Psychophysiology.

[29]  Lars Meyer,et al.  The neural oscillations of speech processing and language comprehension: state of the art and emerging mechanisms , 2018, The European journal of neuroscience.

[30]  Richard J. Davidson,et al.  Electromyogenic artifacts and electroencephalographic inferences revisited , 2011, NeuroImage.

[31]  S. Taulu,et al.  Spatiotemporal signal space separation method for rejecting nearby interference in MEG measurements , 2006, Physics in medicine and biology.

[32]  Terrence J. Sejnowski,et al.  Enhanced detection of artifacts in EEG data using higher-order statistics and independent component analysis , 2007, NeuroImage.

[33]  Manuel Carreiras,et al.  Neocortical activity tracks the hierarchical linguistic structures of self-produced speech during reading aloud , 2020, NeuroImage.

[34]  Riitta Salmelin,et al.  A multimodal spectral approach to characterize rhythm in natural speech. , 2016, The Journal of the Acoustical Society of America.

[35]  Mia Liljeström,et al.  Perceiving and naming actions and objects , 2008, NeuroImage.

[36]  J. Schwartz,et al.  Seeing to hear better: evidence for early audio-visual interactions in speech identification , 2004, Cognition.

[37]  F.-Xavier Alario,et al.  On the cortical dynamics of word production: a review of the MEG evidence , 2016 .

[38]  Robert Oostenveld,et al.  Online and offline tools for head movement compensation in MEG , 2013, NeuroImage.

[39]  Guido Nolte,et al.  Brain Oscillations and Functional Connectivity during Overt Language Production , 2012, Front. Psychology.

[40]  Riitta Salmelin,et al.  Corticomuscular Coherence Is Tuned to the Spontaneous Rhythmicity of Speech at 2–3 Hz , 2012, The Journal of Neuroscience.

[41]  Lesya Y. Ganushchak,et al.  The Use of Electroencephalography in Language Production Research: A Review , 2011, Front. Psychology.