Audiovisual Non-Verbal Dynamic Faces Elicit Converging fMRI and ERP Responses

In an everyday social interaction we automatically integrate another’s facial movements and vocalizations, be they linguistic or otherwise. This requires audiovisual integration of a continual barrage of sensory input—a phenomenon previously well-studied with human audiovisual speech, but not with non-verbal vocalizations. Using both fMRI and ERPs, we assessed neural activity to viewing and listening to an animated female face producing non-verbal, human vocalizations (i.e. coughing, sneezing) under audio-only (AUD), visual-only (VIS) and audiovisual (AV) stimulus conditions, alternating with Rest (R). Underadditive effects occurred in regions dominant for sensory processing, which showed AV activation greater than the dominant modality alone. Right posterior temporal and parietal regions showed an AV maximum in which AV activation was greater than either modality alone, but not greater than the sum of the unisensory conditions. Other frontal and parietal regions showed Common-activation in which AV activation was the same as one or both unisensory conditions. ERP data showed an early superadditive effect (AV > AUD + VIS, no rest), mid-range underadditive effects for auditory N140 and face-sensitive N170, and late AV maximum and common-activation effects. Based on convergence between fMRI and ERP data, we propose a mechanism where a multisensory stimulus may be signaled or facilitated as early as 60 ms and facilitated in sensory-specific regions by increasing processing speed (at N170) and efficiency (decreasing amplitude in auditory and face-sensitive cortical activation and ERPs). Finally, higher-order processes are also altered, but in a more complex fashion.

[1]  T Allison,et al.  ERPS EVOKED BY VIEWING FACIAL MOVEMENTS , 2000, Cognitive neuropsychology.

[2]  G. Pourtois,et al.  Distributed and interactive brain mechanisms during emotion face perception: Evidence from functional neuroimaging , 2007, Neuropsychologia.

[3]  Michael Erb,et al.  Audiovisual integration of emotional signals in voice and face: An event-related fMRI study , 2007, NeuroImage.

[4]  Brigitte Röder,et al.  Multisensory processing in the redundant-target effect: A behavioral and event-related potential study , 2005, Perception & psychophysics.

[5]  R. Kawashima,et al.  Selective Visual and Auditory Attention Toward Utterances—A PET Study , 1999, NeuroImage.

[6]  W. H. Sumby,et al.  Visual contribution to speech intelligibility in noise , 1954 .

[7]  B. Argall,et al.  Unraveling multisensory integration: patchy organization within human STS multisensory cortex , 2004, Nature Neuroscience.

[8]  W. Singer,et al.  Dynamic predictions: Oscillations and synchrony in top–down processing , 2001, Nature Reviews Neuroscience.

[9]  Jeff Miller,et al.  Divided attention: Evidence for coactivation with redundant signals , 1982, Cognitive Psychology.

[10]  S. Iversen,et al.  Detection of Audio-Visual Integration Sites in Humans by Application of Electrophysiological Criteria to the BOLD Effect , 2001, NeuroImage.

[11]  D. Goodin,et al.  Comparison of late components in simultaneously recorded event-related electrical potentials and event-related magnetic fields. , 1996, Electroencephalography and clinical neurophysiology.

[12]  Paul J. Laurienti,et al.  On the use of superadditivity as a metric for characterizing multisensory integration in functional neuroimaging studies , 2005, Experimental Brain Research.

[13]  Y. Sarfati,et al.  Attribution of intentions to others in people with schizophrenia: a non-verbal exploration with comic strips , 1997, Schizophrenia Research.

[14]  John J. Foxe,et al.  The case for feedforward multisensory convergence during early cortical processing , 2005, Neuroreport.

[15]  J. Gore,et al.  A comparison of bound and unbound audio-visual information processing in the human cerebral cortex. , 2002, Brain research. Cognitive brain research.

[16]  John J. Foxe,et al.  Seeing voices: High-density electrical mapping and source-analysis of the multisensory mismatch negativity evoked during the McGurk illusion , 2007, Neuropsychologia.

[17]  P. McGuire,et al.  Silent speechreading in the absence of scanner noise: an event‐related fMRI study , 2000, Neuroreport.

[18]  N. Bolognini,et al.  “Acoustical vision” of below threshold stimuli: interaction among spatially converging audiovisual inputs , 2004, Experimental Brain Research.

[19]  D. V. van Essen,et al.  A Population-Average, Landmark- and Surface-based (PALS) atlas of human cerebral cortex. , 2005, NeuroImage.

[20]  P. Chauvel,et al.  Neuromagnetic source localization of auditory evoked fields and intracerebral evoked potentials: a comparison of data in the same patients , 2001, Clinical Neurophysiology.

[21]  David C. Van Essen,et al.  Application of Information Technology: An Integrated Software Suite for Surface-based Analyses of Cerebral Cortex , 2001, J. Am. Medical Informatics Assoc..

[22]  R W Cox,et al.  AFNI: software for analysis and visualization of functional magnetic resonance neuroimages. , 1996, Computers and biomedical research, an international journal.

[23]  C. R. Michael,et al.  Integration of auditory information in the cat's visual cortex. , 1973, Vision research.

[24]  Michael S. Beauchamp,et al.  Statistical criteria in fMRI studies of multisensory integration , 2005, Neuroinformatics.

[25]  Dorothee J. Chwilla,et al.  When Heuristics Clash with Parsing Routines: ERP Evidence for Conflict Monitoring in Sentence Perception , 2006, Journal of Cognitive Neuroscience.

[26]  Pascal Belin,et al.  Amygdala responses to nonlinguistic emotional vocalizations , 2007, NeuroImage.

[27]  E Macaluso,et al.  Spatial and temporal factors during processing of audiovisual speech: a PET study , 2004, NeuroImage.

[28]  Steven L. Small,et al.  Listening to talking faces: motor cortical activation during speech perception , 2005, NeuroImage.

[29]  D. Perrett,et al.  Integration of form and motion in the anterior superior temporal polysensory area (STPa) of the macaque monkey. , 1996, Journal of neurophysiology.

[30]  Jean Vroomen,et al.  Neural Correlates of Multisensory Integration of Ecologically Valid Audiovisual Events , 2007, Journal of Cognitive Neuroscience.

[31]  Ladan Shams,et al.  Early modulation of visual cortex by sound: an MEG study , 2005, Neuroscience Letters.

[32]  David Poeppel,et al.  Visual speech speeds up the neural processing of auditory speech. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[33]  Erich Schröger,et al.  Prefrontal cortex involvement in preattentive auditory deviance detection: neuroimaging and electrophysiological evidence , 2003, NeuroImage.

[34]  K. Grill-Spector,et al.  Autism and the development of face processing , 2006, Clinical Neuroscience Research.

[35]  A. Puce,et al.  Neuronal oscillations and visual amplification of speech , 2008, Trends in Cognitive Sciences.

[36]  E. Redcay The superior temporal sulcus performs a common function for social and speech perception: Implications for the emergence of autism , 2008, Neuroscience & Biobehavioral Reviews.

[37]  A. Ghazanfar,et al.  Is neocortex essentially multisensory? , 2006, Trends in Cognitive Sciences.

[38]  A. Macleod,et al.  Quantifying the contribution of vision to speech perception in noise. , 1987, British journal of audiology.

[39]  M. Giard,et al.  Auditory-Visual Integration during Multimodal Object Recognition in Humans: A Behavioral and Electrophysiological Study , 1999, Journal of Cognitive Neuroscience.

[40]  R. Knight,et al.  Neural origins of the P300. , 2000, Critical reviews in neurobiology.

[41]  C. Frith,et al.  Social Cognition in Humans , 2007, Current Biology.

[42]  Cheryl M. Capek,et al.  Cortical circuits for silent speechreading in deaf and hearing people , 2008, Neuropsychologia.

[43]  Jean-Philippe Thiran,et al.  Multisensory interactions within human primary cortices revealed by BOLD dynamics. , 2007, Cerebral cortex.

[44]  Jorge Moll,et al.  Moral judgments, emotions and the utilitarian brain , 2007, Trends in Cognitive Sciences.

[45]  John J. Foxe,et al.  Crossmodal binding through neural coherence: implications for multisensory processing , 2008, Trends in Neurosciences.

[46]  B E Walden,et al.  Evaluating the articulation index for auditory-visual consonant recognition. , 1996, The Journal of the Acoustical Society of America.

[47]  J. Eggermont,et al.  Maturation of human central auditory system activity: separating auditory evoked potentials by dipole source modeling , 2002, Clinical Neurophysiology.

[48]  Ryan A. Stevenson,et al.  Superadditive BOLD activation in superior temporal sulcus with threshold non-speech objects , 2007, Experimental Brain Research.

[49]  Gregory McCarthy,et al.  Polysensory interactions along lateral temporal regions evoked by audiovisual speech. , 2003, Cerebral cortex.

[50]  D. Callan,et al.  Giving speech a hand: Gesture modulates activity in auditory cortex during speech perception , 2009, Human brain mapping.

[51]  Christoph Kayser,et al.  Behavioral/systems/cognitive Functional Imaging Reveals Visual Modulation of Specific Fields in Auditory Cortex , 2022 .

[52]  G. Spalletta,et al.  Non‐verbal behaviour deficits in schizophrenia: an ethological study of drug‐free patients , 1998, Acta psychiatrica Scandinavica.

[53]  F. Perrin,et al.  Dissociation of temporal and frontal components in the human auditory N1 wave: a scalp current density and dipole model analysis. , 1994, Electroencephalography and clinical neurophysiology.

[54]  Aina Puce,et al.  Neural responses elicited to face motion and vocalization pairings , 2007, Neuropsychologia.

[55]  FRANK MORRELL,et al.  Visual System's View of Acoustic Space , 1972, Nature.

[56]  D. Perrett,et al.  Neural Representation for the Perception of the Intentionality of Actions , 2000, Brain and Cognition.

[57]  P. McGuire,et al.  Cortical substrates for the perception of face actions: an fMRI study of the specificity of activation for seen speech and for meaningless lower-face acts (gurning). , 2001, Brain research. Cognitive brain research.

[58]  A. Fort,et al.  Bimodal speech: early suppressive visual effects in human auditory cortex , 2004, The European journal of neuroscience.

[59]  Helen Tager-Flusberg,et al.  Language Assessment and Development in Toddlers with Autism Spectrum Disorders , 2008, Journal of autism and developmental disorders.

[60]  Joost X. Maier,et al.  Multisensory Integration of Dynamic Faces and Voices in Rhesus Monkey Auditory Cortex , 2005 .

[61]  Umberto Castiello,et al.  The human temporal lobe integrates facial form and motion: evidence from fMRI and ERP studies , 2003, NeuroImage.