Person identification through faces and voices: An ERP study

Different models have been proposed to explain how identity is extracted from faces and voices and how these two sensory systems interact. The neural loci of audio-visual interactions have been studied using neuroimaging techniques; however, the time course of these interactions is not well established. Here, we use event related potentials (ERPs) to study the temporal dynamics of the interaction of face and voice processing modules. We presented to the subjects either faces alone (F), voices alone (V) or faces and voices together (FV) in a familiarity detection task. Responses obtained for FV were compared with the sum of the responses obtained for F plus responses for V, for familiar and unfamiliar stimuli. This comparison shows differences in amplitude for different latencies, indicating cross-modal interactions. For unfamiliar stimuli, this interaction began very early (around 200ms) and was restricted to the time window corresponding to the face N170 component. For familiar stimuli, the interaction was longer, began earlier and remained until after the N170 component. These results indicate that the interaction between faces and voices occurs from the initial stages of processing and continues as the person identification process goes on. This study is the first electrophysiological evidence of cross-modal interaction of faces and voices during the recognition of acquaintances' identities. It suggests that the assessment of person familiarity can result in direct information sharing between voice and face sensory modules from the early processing stages, before access to the person identity nodes.

[1]  Y. Benjamini,et al.  Controlling the false discovery rate: a practical and powerful approach to multiple testing , 1995 .

[2]  Mark H Johnson,et al.  Development of face-sensitive event-related potentials during infancy: a review. , 2003, International journal of psychophysiology : official journal of the International Organization of Psychophysiology.

[3]  Ryan A. Stevenson,et al.  Audiovisual integration in human superior temporal sulcus: Inverse effectiveness and the neural processing of speech and object recognition , 2009, NeuroImage.

[4]  Bruno Rossion,et al.  The functionally defined right occipital and fusiform “face areas” discriminate novel from visually familiar faces , 2003, NeuroImage.

[5]  N. Kanwisher,et al.  The Fusiform Face Area: A Module in Human Extrastriate Cortex Specialized for Face Perception , 1997, The Journal of Neuroscience.

[6]  K. G. Munhall,et al.  Audiovisual Integration of Speech in a Bistable Illusion , 2009, Current Biology.

[7]  S. Bentin,et al.  Processing specificity for human voice stimuli: electrophysiological evidence , 2001, Neuroreport.

[8]  Andreas Kleinschmidt,et al.  Interaction of Face and Voice Areas during Speaker Recognition , 2005, Journal of Cognitive Neuroscience.

[9]  Dylan M. Jones,et al.  Intra- and inter-modal repetition priming of familiar faces and voices. , 1997, British journal of psychology.

[10]  M. Hallett,et al.  Neural Correlates of Auditory–Visual Stimulus Onset Asynchrony Detection , 2001, The Journal of Neuroscience.

[11]  B. Efron Large-Scale Simultaneous Hypothesis Testing , 2004 .

[12]  J. Polich Updating P300: An integrative theory of P3a and P3b , 2007, Clinical Neurophysiology.

[13]  Shlomo Bentin,et al.  Neural sensitivity to human voices: ERP evidence of task and attentional influences. , 2003, Psychophysiology.

[14]  L. Pessoa To what extent are emotional visual stimuli processed without attention and awareness? , 2005, Current Opinion in Neurobiology.

[15]  A. Giraud,et al.  Implicit Multisensory Associations Influence Voice Recognition , 2006, PLoS biology.

[16]  Susan M. Hughes,et al.  The processing of auditory and visual recognition of self-stimuli , 2010, Consciousness and Cognition.

[17]  M. Tarr,et al.  The Fusiform Face Area is Part of a Network that Processes Faces at the Individual Level , 2000, Journal of Cognitive Neuroscience.

[18]  A. Burton,et al.  Event-related brain potential evidence for a response of inferior temporal cortex to familiar face repetitions. , 2002, Brain research. Cognitive brain research.

[19]  M. Eimer Effects of face inversion on the structural encoding and recognition of faces. Evidence from event-related brain potentials. , 2000, Brain research. Cognitive brain research.

[20]  S. Campanella,et al.  Integrating face and voice in person perception , 2007, Trends in Cognitive Sciences.

[21]  S. Schweinberger,et al.  Hearing Facial Identities , 2007, Quarterly journal of experimental psychology.

[22]  J. Haxby,et al.  The distributed human neural system for face perception , 2000, Trends in Cognitive Sciences.

[23]  N. Sagiv,et al.  Structural Encoding of Human and Schematic Faces: Holistic and Part-Based Processes , 2001, Journal of Cognitive Neuroscience.

[24]  Bruno Debruille,et al.  Brain potentials reveal covert facial recognition in prosopagnosia , 1989, Neuropsychologia.

[25]  S. Schweinberger,et al.  The role of audiovisual asynchrony in person recognition , 2010, Quarterly journal of experimental psychology.

[26]  S A Hillyard,et al.  An analysis of audio-visual crossmodal integration by means of event-related potential (ERP) recordings. , 2002, Brain research. Cognitive brain research.

[27]  S. Bentin,et al.  Domain specificity versus expertise: factors influencing distinct processing of faces , 2002, Cognition.

[28]  Arthur W. Toga,et al.  A Probabilistic Atlas of the Human Brain: Theory and Rationale for Its Development The International Consortium for Brain Mapping (ICBM) , 1995, NeuroImage.

[29]  M. Eimer Event-related brain potentials distinguish processing stages involved in face perception and recognition , 2000, Clinical Neurophysiology.

[30]  R. Blair,et al.  An alternative method for significance testing of waveform difference potentials. , 1993, Psychophysiology.

[31]  John J. Foxe,et al.  The case for feedforward multisensory convergence during early cortical processing , 2005, Neuroreport.

[32]  A. Burton,et al.  N250r: a face-selective brain response to stimulus repetitions , 2004, Neuroreport.

[33]  Lucy J. Miller,et al.  Multisensory integration in children: A preliminary ERP study , 2008, Brain Research.

[34]  D. Regan,et al.  Evoked potentials in vision research 1961–1986 , 1986, Vision Research.

[35]  M. Giard,et al.  Auditory-Visual Integration during Multimodal Object Recognition in Humans: A Behavioral and Electrophysiological Study , 1999, Journal of Cognitive Neuroscience.

[36]  John J. Foxe,et al.  Multisensory auditory-visual interactions during early sensory processing in humans: a high-density electrical mapping study. , 2002, Brain research. Cognitive brain research.

[37]  R. Hari,et al.  Seeing speech: visual information from lip movements modifies activity in the human auditory cortex , 1991, Neuroscience Letters.

[38]  R. Zatorre,et al.  Voice-selective areas in human auditory cortex , 2000, Nature.

[39]  Stefan J Kiebel,et al.  How the Human Brain Recognizes Speech in the Context of Changing Speakers , 2010, The Journal of Neuroscience.

[40]  Joel L. Voss,et al.  Validating neural correlates of familiarity , 2007, Trends in Cognitive Sciences.

[41]  R. Verleger On the utility of P3 latency as an index of mental chronometry. , 1997, Psychophysiology.

[42]  Denise A. Minnebusch,et al.  A bilateral occipitotemporal network mediates face perception , 2009, Behavioural Brain Research.

[43]  G. McCarthy,et al.  Augmenting mental chronometry: the P300 as a measure of stimulus evaluation time. , 1977, Science.

[44]  Sean J. O'Connor,et al.  P300 topography of amplitude/latency correlations , 2005, Brain Topography.

[45]  M. Crommelinck,et al.  When audition alters vision: an event-related potential study of the cross-modal interactions between faces and voices , 2004, Neuroscience Letters.

[46]  Margot J. Taylor,et al.  Spatio temporal dynamics of face recognition. , 2008, Cerebral cortex.

[47]  Michael S. Beauchamp,et al.  Statistical criteria in fMRI studies of multisensory integration , 2005, Neuroinformatics.

[48]  David Poeppel,et al.  Visual speech speeds up the neural processing of auditory speech. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[49]  L. Balázs,et al.  Event-related potentials and audiovisual stimuli: multimodal interactions , 2001, Neuroreport.

[50]  K. Schlingensiepen,et al.  Sequence‐specific impairment of learning by c-jun antisense oligonucleotides , 1994, Neuroreport.

[51]  M. Tarr,et al.  The N170 occipito‐temporal component is delayed and enhanced to inverted faces but not to inverted objects: an electrophysiological account of face‐specific processes in the human brain , 2000, Neuroreport.

[52]  Leslie G. Ungerleider,et al.  The Effect of Face Inversion on Activity in Human Neural Systems for Face and Object Perception , 1999, Neuron.

[53]  E. Phelps Emotion and cognition: insights from studies of the human amygdala. , 2006, Annual review of psychology.

[54]  F. Lopera,et al.  Brain potentials reflect residual face processing in a case of prosopagnosia , 2004, Cognitive neuropsychology.

[55]  A. Damasio The Feeling of What Happens: Body and Emotion in the Making of Consciousness , 1999 .

[56]  G. Pourtois,et al.  Perception of Facial Expressions and Voices and of their Combination in the Human Brain , 2005, Cortex.

[57]  J. Shedden,et al.  Automatic face identity encoding at the N170 , 2006, Vision Research.

[58]  Eduardo Martínez-Montes,et al.  False discovery rate and permutation test: An evaluation in ERP data analysis , 2010, Statistics in medicine.

[59]  David I. Perrett,et al.  Event-related potentials and the matching of familiar and unfamiliar faces , 1988, Neuropsychologia.

[60]  L. Deouell,et al.  STRUCTURAL ENCODING AND IDENTIFICATION IN FACE PROCESSING: ERP EVIDENCE FOR SEPARATE MECHANISMS , 2000, Cognitive neuropsychology.

[61]  Marcia Grabowecky,et al.  Electrophysiological Correlates of Recollecting Faces of Known and Unknown Individuals , 2000, NeuroImage.

[62]  T. Allison,et al.  Electrophysiological Studies of Face Perception in Humans , 1996, Journal of Cognitive Neuroscience.

[63]  P. Belin,et al.  Thinking the voice: neural correlates of voice perception , 2004, Trends in Cognitive Sciences.

[64]  H Begleiter,et al.  Event-related brain potentials differentiate priming and recognition to familiar and unfamiliar faces. , 1995, Electroencephalography and clinical neurophysiology.

[65]  Stefan R. Schweinberger,et al.  N250r ERP repetition effects from distractor faces when attending to another face under load: Evidence for a face attention resource , 2009, Brain Research.

[66]  H. McGurk,et al.  Hearing lips and seeing voices , 1976, Nature.

[67]  Anne-Lise Giraud,et al.  Voice recognition and cross-modal responses to familiar speakers' voices in prosopagnosia. , 2006, Cerebral cortex.

[68]  C. Nelson,et al.  Event-related potentials in year-old infants: relations with emotionality and cortisol. , 1994, Child development.

[69]  John J. Foxe,et al.  Multisensory auditory-somatosensory interactions in early cortical processing revealed by high-density electrical mapping. , 2000, Brain research. Cognitive brain research.

[70]  Luc H. Arnal,et al.  Dual Neural Routing of Visual Facilitation in Speech Processing , 2009, The Journal of Neuroscience.

[71]  Werner Sommer,et al.  Repetition priming and associative priming of face recognition: Evidence from event-related potentials. , 1995 .

[72]  Michael Erb,et al.  Audiovisual integration of emotional signals in voice and face: An event-related fMRI study , 2007, NeuroImage.

[73]  A. Young,et al.  Understanding face recognition. , 1986, British journal of psychology.

[74]  Thomas E. Nichols,et al.  Nonparametric permutation tests for functional neuroimaging: A primer with examples , 2002, Human brain mapping.

[75]  B. Rossion,et al.  Task modulation of brain activity related to familiar and unfamiliar face processing: an ERP study , 1999, Clinical Neurophysiology.

[76]  Pascal Belin,et al.  Electrophysiological evidence for an early processing of human voices , 2009, BMC Neuroscience.

[77]  Maria A. Bobes,et al.  Brain potentials reflect access to visual and emotional memories for faces , 2007, Biological Psychology.

[78]  T W Picton,et al.  The P300 Wave of the Human Event‐Related Potential , 1992, Journal of clinical neurophysiology : official publication of the American Electroencephalographic Society.

[79]  R. Biscay,et al.  Testing topographic differences between event related brain potentials by using non-parametric combinations of permutation tests. , 1997, Electroencephalography and clinical neurophysiology.