Temporal factors in the electrophysiological markers of audiovisual speech integration
暂无分享,去创建一个
[1] E Macaluso,et al. Spatial and temporal factors during processing of audiovisual speech: a PET study , 2004, NeuroImage.
[2] David Poeppel,et al. Visual speech speeds up the neural processing of auditory speech. , 2005, Proceedings of the National Academy of Sciences of the United States of America.
[3] J. Navarra,et al. Assessing automaticity in audiovisual speech integration: evidence from the speeded classification task , 2004, Cognition.
[4] M. Sams,et al. Electrophysiological indicators of phonetic and non-phonetic multisensory interactions during audiovisual speech perception. , 2003, Brain research. Cognitive brain research.
[5] P. McGuire,et al. Silent speechreading in the absence of scanner noise: an event‐related fMRI study , 2000, Neuroreport.
[6] Tom C T Yin,et al. Bimodal Interactions in the Superior Colliculus of the Behaving Cat , 2002, The Journal of Neuroscience.
[7] P. Gribble,et al. Temporal constraints on the McGurk effect , 1996, Perception & psychophysics.
[8] M. Hyde,et al. The N1 response and its applications. , 1997, Audiology & neuro-otology.
[9] A. Faulkner,et al. Adaptation by normal listeners to upward spectral shifts of speech: implications for cochlear implants. , 1999, The Journal of the Acoustical Society of America.
[10] L. Rosenblum,et al. An audiovisual test of kinematic primitives for visual speech perception. , 1996, Journal of experimental psychology. Human perception and performance.
[11] H. McGurk,et al. Hearing lips and seeing voices , 1976, Nature.
[12] L. Braida. Crossmodal Integration in the Identification of Consonant Segments , 1991, The Quarterly journal of experimental psychology. A, Human experimental psychology.
[13] David B Pisoni,et al. Specification of cross-modal source information in isolated kinematic displays of speech. , 2004, The Journal of the Acoustical Society of America.
[14] K. Green. The Use of Auditory and Visual Information in Phonetic Perception , 1996 .
[15] Y. Yen,et al. Deactivation of Sensory-Specific Cortex by Cross-Modal Stimuli , 2002, Journal of Cognitive Neuroscience.
[16] G. Plant. Perceiving Talking Faces: From Speech Perception to a Behavioral Principle , 1999 .
[17] R. Campbell,et al. Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex , 2000, Current Biology.
[18] A. Fort,et al. Bimodal speech: early suppressive visual effects in human auditory cortex , 2004, The European journal of neuroscience.
[19] Q. Summerfield,et al. Intermodal timing relations and audio-visual speech recognition by normal-hearing adults. , 1985, The Journal of the Acoustical Society of America.
[20] D. Reisberg,et al. Easy to hear but hard to understand: A lip-reading advantage with intact auditory stimuli. , 1987 .
[21] David Taylor. Hearing by Eye: The Psychology of Lip-Reading , 1988 .
[22] A. Fort,et al. Is the auditory sensory memory sensitive to visual information? , 2005, Experimental Brain Research.
[23] J. Schwartz,et al. Seeing to hear better: evidence for early audio-visual interactions in speech identification , 2004, Cognition.
[24] W. H. Sumby,et al. Visual contribution to speech intelligibility in noise , 1954 .
[25] I. Winkler,et al. The concept of auditory stimulus representation in cognitive neuroscience. , 1999, Psychological bulletin.
[26] Jeffery A. Jones,et al. Multisensory Integration Sites Identified by Perception of Spatial Wavelet Filtered Visual Speech Gesture Information , 2004, Journal of Cognitive Neuroscience.