A Visual Cortical Network for Deriving Phonological Information from Intelligible Lip Movements
暂无分享,去创建一个
Elisa Leonardelli | Olivier Collignon | Anne Hauswald | Nathan Weisz | Chrysa D. Lithari | N. Weisz | O. Collignon | A. Hauswald | E. Leonardelli | C. Lithari | Chrysa Lithari | Elisa Leonardelli
[1] Jeffrey F. Cohn,et al. Robust Lip Tracking by Combining Shape, Color and Motion , 2007 .
[2] Robin A A Ince,et al. Irregular Speech Rate Dissociates Auditory Cortical Entrainment, Evoked Responses, and Frontal Alpha , 2015, The Journal of Neuroscience.
[3] W. Drongelen,et al. Localization of brain electrical activity via linearly constrained minimum variance spatial filtering , 1997, IEEE Transactions on Biomedical Engineering.
[4] Mingzhou Ding,et al. Analyzing information flow in brain networks with nonparametric Granger causality , 2008, NeuroImage.
[5] Whitney M. Weikum,et al. Visual Language Discrimination in Infancy , 2007, Science.
[6] David Poeppel,et al. Cortical oscillations and speech processing: emerging computational principles and operations , 2012, Nature Neuroscience.
[7] M. Sams,et al. Primary auditory cortex activation by visual speech: an fMRI study at 3 T , 2005, Neuroreport.
[8] Zachary M. Smith,et al. Chimaeric sounds reveal dichotomies in auditory perception , 2002, Nature.
[9] A. Ghazanfar,et al. Is neocortex essentially multisensory? , 2006, Trends in Cognitive Sciences.
[10] A. Giraud,et al. Faster phonological processing and right occipito-temporal coupling in deaf adults signal poor cochlear implant outcome , 2017, Nature Communications.
[11] Asif A. Ghazanfar,et al. The Natural Statistics of Audiovisual Speech , 2009, PLoS Comput. Biol..
[12] Josef P. Rauschecker,et al. Ventral and dorsal streams in the evolution of speech and language , 2011, Front. Evol. Neurosci..
[13] Gregor Thut,et al. Lip movements entrain the observers’ low-frequency brain oscillations to facilitate speech intelligibility , 2016, eLife.
[14] Stefano Panzeri,et al. Contributions of local speech encoding and functional connectivity to audio-visual speech perception , 2017, eLife.
[15] E. Liebenthal,et al. Neural pathways for visual speech perception , 2014, Front. Neurosci..
[16] Ulf Andersson,et al. Bottom-Up Driven Speechreading in a Speechreading Expert: The Case of AA (JK023) , 2005, Ear and hearing.
[17] D H Brainard,et al. The Psychophysics Toolbox. , 1997, Spatial vision.
[18] Nancy Tye-Murray,et al. Effects of Context Type on Lipreading and Listening Performance and Implications for Sentence Processing. , 2015, Journal of speech, language, and hearing research : JSLHR.
[19] H. Kennedy,et al. Alpha-Beta and Gamma Rhythms Subserve Feedback and Feedforward Influences among Human Visual Cortical Areas , 2016, Neuron.
[20] P. Schyns,et al. Speech Rhythms and Multiplexed Oscillatory Sensory Coding in the Human Brain , 2013, PLoS biology.
[21] J. Rauschecker. Cortical processing of complex sounds , 1998, Current Opinion in Neurobiology.
[22] Karl J. Friston,et al. Canonical Source Reconstruction for MEG , 2007, Comput. Intell. Neurosci..
[23] Whitney M. Weikum,et al. Visual language discrimination , 2008 .
[24] D. Lewkowicz. Perception of auditory-visual temporal synchrony in human infants. , 1996, Journal of experimental psychology. Human perception and performance.
[25] H. Kennedy,et al. Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels , 2014, Neuron.
[26] Michael S. Beauchamp,et al. A Double Dissociation between Anterior and Posterior Superior Temporal Gyrus for Processing Audiovisual Speech Demonstrated by Electrocorticography , 2017, Journal of Cognitive Neuroscience.
[27] Colin Humphries,et al. Role of left posterior superior temporal gyrus in phonological processing for speech perception and production , 2001, Cogn. Sci..
[28] John J. Foxe,et al. Neural responses to uninterrupted natural speech can be extracted with precise temporal resolution , 2010, The European journal of neuroscience.
[29] D. Lewkowicz,et al. Multisensory Processes: A Balancing Act across the Lifespan , 2016, Trends in Neurosciences.
[30] Ruth Campbell,et al. The processing of audio-visual speech: empirical and neural bases , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.
[31] E. Bullmore,et al. Activation of auditory cortex during silent lipreading. , 1997, Science.
[32] D. Poeppel,et al. Auditory Cortex Tracks Both Auditory and Visual Stimulus Dynamics Using Low-Frequency Neuronal Phase Modulation , 2010, PLoS biology.
[33] David C. Van Essen,et al. Application of Information Technology: An Integrated Software Suite for Surface-based Analyses of Cerebral Cortex , 2001, J. Am. Medical Informatics Assoc..
[34] R. Hari,et al. Seeing speech: visual information from lip movements modifies activity in the human auditory cortex , 1991, Neuroscience Letters.
[35] J. Rauschecker,et al. Processing of complex sounds in the macaque nonprimary auditory cortex. , 1995, Science.
[36] D. Lewkowicz,et al. Infants deploy selective attention to the mouth of a talking face when learning speech , 2012, Proceedings of the National Academy of Sciences.
[37] Oded Ghitza,et al. On the Role of Theta-Driven Syllabic Parsing in Decoding Speech: Intelligibility of Speech with a Manipulated Modulation Spectrum , 2012, Front. Psychology.
[38] R. Oostenveld,et al. Nonparametric statistical testing of EEG- and MEG-data , 2007, Journal of Neuroscience Methods.
[39] Robin A. A. Ince,et al. Frontal Top-Down Signals Increase Coupling of Auditory Low-Frequency Oscillations to Continuous Speech in Human Listeners , 2015, Current Biology.
[40] G. Nolte. The magnetic lead field theorem in the quasi-static approximation and its use for magnetoencephalography forward calculation in realistic volume conductors. , 2003, Physics in medicine and biology.
[41] N. Tzourio-Mazoyer,et al. Automated Anatomical Labeling of Activations in SPM Using a Macroscopic Anatomical Parcellation of the MNI MRI Single-Subject Brain , 2002, NeuroImage.
[42] Robert Oostenveld,et al. FieldTrip: Open Source Software for Advanced Analysis of MEG, EEG, and Invasive Electrophysiological Data , 2010, Comput. Intell. Neurosci..
[43] Edmund C. Lalor,et al. Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent Lipreading , 2017, Front. Hum. Neurosci..
[44] D. Poeppel,et al. The cortical organization of speech processing , 2007, Nature Reviews Neuroscience.
[45] Josef P. Rauschecker,et al. Where, When, and How: Are they all sensorimotor? Towards a unified view of the dorsal pathway in vision and audition , 2018, Cortex.
[46] H. Nusbaum,et al. Visual cortex entrains to sign language , 2017, Proceedings of the National Academy of Sciences.