A Visual Cortical Network for Deriving Phonological Information from Intelligible Lip Movements

Summary Successful lip-reading requires a mapping from visual to phonological information [1]. Recently, visual and motor cortices have been implicated in tracking lip movements (e.g., [2]). It remains unclear, however, whether visuo-phonological mapping occurs already at the level of the visual cortex–that is, whether this structure tracks the acoustic signal in a functionally relevant manner. To elucidate this, we investigated how the cortex tracks (i.e., entrains to) absent acoustic speech signals carried by silent lip movements. Crucially, we contrasted the entrainment to unheard forward (intelligible) and backward (unintelligible) acoustic speech. We observed that the visual cortex exhibited stronger entrainment to the unheard forward acoustic speech envelope compared to the unheard backward acoustic speech envelope. Supporting the notion of a visuo-phonological mapping process, this forward-backward difference of occipital entrainment was not present for actually observed lip movements. Importantly, the respective occipital region received more top-down input, especially from left premotor, primary motor, and somatosensory regions and, to a lesser extent, also from posterior temporal cortex. Strikingly, across participants, the extent of top-down modulation of the visual cortex stemming from these regions partially correlated with the strength of entrainment to absent acoustic forward speech envelope, but not to present forward lip movements. Our findings demonstrate that a distributed cortical network, including key dorsal stream auditory regions [3, 4, 5], influences how the visual cortex shows sensitivity to the intelligibility of speech while tracking silent lip movements.

[1]  Jeffrey F. Cohn,et al.  Robust Lip Tracking by Combining Shape, Color and Motion , 2007 .

[2]  Robin A A Ince,et al.  Irregular Speech Rate Dissociates Auditory Cortical Entrainment, Evoked Responses, and Frontal Alpha , 2015, The Journal of Neuroscience.

[3]  W. Drongelen,et al.  Localization of brain electrical activity via linearly constrained minimum variance spatial filtering , 1997, IEEE Transactions on Biomedical Engineering.

[4]  Mingzhou Ding,et al.  Analyzing information flow in brain networks with nonparametric Granger causality , 2008, NeuroImage.

[5]  Whitney M. Weikum,et al.  Visual Language Discrimination in Infancy , 2007, Science.

[6]  David Poeppel,et al.  Cortical oscillations and speech processing: emerging computational principles and operations , 2012, Nature Neuroscience.

[7]  M. Sams,et al.  Primary auditory cortex activation by visual speech: an fMRI study at 3 T , 2005, Neuroreport.

[8]  Zachary M. Smith,et al.  Chimaeric sounds reveal dichotomies in auditory perception , 2002, Nature.

[9]  A. Ghazanfar,et al.  Is neocortex essentially multisensory? , 2006, Trends in Cognitive Sciences.

[10]  A. Giraud,et al.  Faster phonological processing and right occipito-temporal coupling in deaf adults signal poor cochlear implant outcome , 2017, Nature Communications.

[11]  Asif A. Ghazanfar,et al.  The Natural Statistics of Audiovisual Speech , 2009, PLoS Comput. Biol..

[12]  Josef P. Rauschecker,et al.  Ventral and dorsal streams in the evolution of speech and language , 2011, Front. Evol. Neurosci..

[13]  Gregor Thut,et al.  Lip movements entrain the observers’ low-frequency brain oscillations to facilitate speech intelligibility , 2016, eLife.

[14]  Stefano Panzeri,et al.  Contributions of local speech encoding and functional connectivity to audio-visual speech perception , 2017, eLife.

[15]  E. Liebenthal,et al.  Neural pathways for visual speech perception , 2014, Front. Neurosci..

[16]  Ulf Andersson,et al.  Bottom-Up Driven Speechreading in a Speechreading Expert: The Case of AA (JK023) , 2005, Ear and hearing.

[17]  D H Brainard,et al.  The Psychophysics Toolbox. , 1997, Spatial vision.

[18]  Nancy Tye-Murray,et al.  Effects of Context Type on Lipreading and Listening Performance and Implications for Sentence Processing. , 2015, Journal of speech, language, and hearing research : JSLHR.

[19]  H. Kennedy,et al.  Alpha-Beta and Gamma Rhythms Subserve Feedback and Feedforward Influences among Human Visual Cortical Areas , 2016, Neuron.

[20]  P. Schyns,et al.  Speech Rhythms and Multiplexed Oscillatory Sensory Coding in the Human Brain , 2013, PLoS biology.

[21]  J. Rauschecker Cortical processing of complex sounds , 1998, Current Opinion in Neurobiology.

[22]  Karl J. Friston,et al.  Canonical Source Reconstruction for MEG , 2007, Comput. Intell. Neurosci..

[23]  Whitney M. Weikum,et al.  Visual language discrimination , 2008 .

[24]  D. Lewkowicz Perception of auditory-visual temporal synchrony in human infants. , 1996, Journal of experimental psychology. Human perception and performance.

[25]  H. Kennedy,et al.  Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels , 2014, Neuron.

[26]  Michael S. Beauchamp,et al.  A Double Dissociation between Anterior and Posterior Superior Temporal Gyrus for Processing Audiovisual Speech Demonstrated by Electrocorticography , 2017, Journal of Cognitive Neuroscience.

[27]  Colin Humphries,et al.  Role of left posterior superior temporal gyrus in phonological processing for speech perception and production , 2001, Cogn. Sci..

[28]  John J. Foxe,et al.  Neural responses to uninterrupted natural speech can be extracted with precise temporal resolution , 2010, The European journal of neuroscience.

[29]  D. Lewkowicz,et al.  Multisensory Processes: A Balancing Act across the Lifespan , 2016, Trends in Neurosciences.

[30]  Ruth Campbell,et al.  The processing of audio-visual speech: empirical and neural bases , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.

[31]  E. Bullmore,et al.  Activation of auditory cortex during silent lipreading. , 1997, Science.

[32]  D. Poeppel,et al.  Auditory Cortex Tracks Both Auditory and Visual Stimulus Dynamics Using Low-Frequency Neuronal Phase Modulation , 2010, PLoS biology.

[33]  David C. Van Essen,et al.  Application of Information Technology: An Integrated Software Suite for Surface-based Analyses of Cerebral Cortex , 2001, J. Am. Medical Informatics Assoc..

[34]  R. Hari,et al.  Seeing speech: visual information from lip movements modifies activity in the human auditory cortex , 1991, Neuroscience Letters.

[35]  J. Rauschecker,et al.  Processing of complex sounds in the macaque nonprimary auditory cortex. , 1995, Science.

[36]  D. Lewkowicz,et al.  Infants deploy selective attention to the mouth of a talking face when learning speech , 2012, Proceedings of the National Academy of Sciences.

[37]  Oded Ghitza,et al.  On the Role of Theta-Driven Syllabic Parsing in Decoding Speech: Intelligibility of Speech with a Manipulated Modulation Spectrum , 2012, Front. Psychology.

[38]  R. Oostenveld,et al.  Nonparametric statistical testing of EEG- and MEG-data , 2007, Journal of Neuroscience Methods.

[39]  Robin A. A. Ince,et al.  Frontal Top-Down Signals Increase Coupling of Auditory Low-Frequency Oscillations to Continuous Speech in Human Listeners , 2015, Current Biology.

[40]  G. Nolte The magnetic lead field theorem in the quasi-static approximation and its use for magnetoencephalography forward calculation in realistic volume conductors. , 2003, Physics in medicine and biology.

[41]  N. Tzourio-Mazoyer,et al.  Automated Anatomical Labeling of Activations in SPM Using a Macroscopic Anatomical Parcellation of the MNI MRI Single-Subject Brain , 2002, NeuroImage.

[42]  Robert Oostenveld,et al.  FieldTrip: Open Source Software for Advanced Analysis of MEG, EEG, and Invasive Electrophysiological Data , 2010, Comput. Intell. Neurosci..

[43]  Edmund C. Lalor,et al.  Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent Lipreading , 2017, Front. Hum. Neurosci..

[44]  D. Poeppel,et al.  The cortical organization of speech processing , 2007, Nature Reviews Neuroscience.

[45]  Josef P. Rauschecker,et al.  Where, When, and How: Are they all sensorimotor? Towards a unified view of the dorsal pathway in vision and audition , 2018, Cortex.

[46]  H. Nusbaum,et al.  Visual cortex entrains to sign language , 2017, Proceedings of the National Academy of Sciences.