Functional selectivity for face processing in the temporal voice area of early deaf individuals

Significance Here, we show that deaf individuals activate a specific and discrete subregion of the temporal cortex, typically selective to voices in hearing people, for visual face processing. This reorganized “voice” region participates in face identity processing and responds selectively to faces early in time, suggesting that this area becomes an integral part of the face network in early deaf individuals. Observing that face processing selectively colonizes a region of the hearing brain that is functionally related to identity processing evidences the intrinsic constraints imposed to cross-modal plasticity. Our work therefore supports the view that, even if brain components modify their sensory tuning in case of deprivation, they maintain a relation to the computational structure of the problems they solve. Brain systems supporting face and voice processing both contribute to the extraction of important information for social interaction (e.g., person identity). How does the brain reorganize when one of these channels is absent? Here, we explore this question by combining behavioral and multimodal neuroimaging measures (magneto-encephalography and functional imaging) in a group of early deaf humans. We show enhanced selective neural response for faces and for individual face coding in a specific region of the auditory cortex that is typically specialized for voice perception in hearing individuals. In this region, selectivity to face signals emerges early in the visual processing hierarchy, shortly after typical face-selective responses in the ventral visual pathway. Functional and effective connectivity analyses suggest reorganization in long-range connections from early visual areas to the face-selective temporal area in individuals with early and profound deafness. Altogether, these observations demonstrate that regions that typically specialize for voice processing in the hearing brain preferentially reorganize for face processing in born-deaf people. Our results support the idea that cross-modal plasticity in the case of early sensory deprivation relates to the original functional specialization of the reorganized brain regions.

[1]  Á. Pascual-Leone,et al.  The metamodal organization of the brain. , 2001, Progress in brain research.

[2]  S. Lomber,et al.  Cross-modal plasticity in specific auditory cortices underlies visual compensations in the deaf , 2010, Nature Neuroscience.

[3]  Amir Amedi,et al.  A Ventral Visual Stream Reading Center Independent of Visual Experience , 2012, Current Biology.

[4]  Jon Driver,et al.  Effective Connectivity from Early Visual Cortex to Posterior Occipitotemporal Face Areas Supports Face Selectivity and Predicts Developmental Prosopagnosia , 2016, The Journal of Neuroscience.

[5]  Pascal Belin,et al.  Norm-Based Coding of Voice Identity in Human Auditory Cortex , 2013, Current Biology.

[6]  Murray Mm,et al.  Visual Abilities in Individuals with Profound Deafness A Critical Review -- The Neural Bases of Multisensory Processes , 2012 .

[7]  M. Lassonde,et al.  Cross-modal plasticity for the spatial processing of sounds in visually deprived subjects , 2008, Experimental Brain Research.

[8]  Robert J. Zatorre,et al.  Reorganization of Auditory Cortex in Early-deaf People: Functional Connectivity and Relationship to Hearing Aid Use , 2015, Journal of Cognitive Neuroscience.

[9]  A. Giraud,et al.  Cortical activity at rest predicts cochlear implantation outcome. , 2007, Cerebral cortex.

[10]  N. Kanwisher,et al.  The Fusiform Face Area: A Module in Human Extrastriate Cortex Specialized for Face Perception , 1997, The Journal of Neuroscience.

[11]  Ione Fine,et al.  Visual stimuli activate auditory cortex in the deaf , 2001, Nature Neuroscience.

[12]  Krzysztof J. Gorgolewski,et al.  The human voice areas: Spatial organization and inter-individual variability in temporal and extra-temporal cortices , 2015, NeuroImage.

[13]  N. Logothetis,et al.  Neuroperception: Facial expressions linked to monkey calls , 2003, Nature.

[14]  P. Barone,et al.  Evolution of crossmodal reorganization of the voice area in cochlear‐implanted deaf patients , 2012, Human brain mapping.

[15]  M. Nachman,et al.  Morphological and population genomic evidence that human faces have evolved to signal individual identity , 2014, Nature Communications.

[16]  Marcin Szwed,et al.  Task-specific reorganization of the auditory cortex in deaf humans , 2017, Proceedings of the National Academy of Sciences.

[17]  G. Vandewalle,et al.  Functional specialization for auditory–spatial processing in the occipital cortex of congenitally blind humans , 2011, Proceedings of the National Academy of Sciences.

[18]  N. Logothetis,et al.  Auditory and Visual Modulation of Temporal Lobe Neurons in Voice-Sensitive and Association Cortices , 2014, The Journal of Neuroscience.

[19]  G. Vandewalle,et al.  Impact of blindness onset on the functional organization and the connectivity of the occipital cortex. , 2013, Brain : a journal of neurology.

[20]  Maarten De Vos,et al.  Cross-modal reorganization in cochlear implant users: Auditory cortex contributes to visual face processing , 2015, NeuroImage.

[21]  Karl J. Friston,et al.  Dynamic causal models of neural system dynamics: current state and future extensions , 2007, Journal of Biosciences.

[22]  A. Norcia,et al.  The 6Hz fundamental stimulation frequency rate for individual face discrimination in the right occipito-temporal cortex , 2013, Neuropsychologia.

[23]  Eric Halgren,et al.  Signed Words in the Congenitally Deaf Evoke Typical Late Lexicosemantic Responses with No Early Visual Responses in Left Superior Temporal Cortex , 2012, The Journal of Neuroscience.

[24]  Josef Parvizi,et al.  Corresponding ECoG and fMRI category-selective signals in human ventral temporal cortex , 2016, Neuropsychologia.

[25]  Karl J. Friston,et al.  Psychophysiological and Modulatory Interactions in Neuroimaging , 1997, NeuroImage.

[26]  Karl J. Friston,et al.  Bayesian model selection for group studies , 2009, NeuroImage.

[27]  R. C. Oldfield The assessment and analysis of handedness: the Edinburgh inventory. , 1971, Neuropsychologia.

[28]  G. Yovel,et al.  A unified coding strategy for processing faces and voices , 2013, Trends in Cognitive Sciences.

[29]  Chunshui Yu,et al.  Cross-modal activation of auditory regions during visuo-spatial working memory in early deafness. , 2015, Brain : a journal of neurology.

[30]  H. Neville,et al.  Altered Cross-Modal Processing in the Primary Auditory Cortex of Congenitally Deaf Adults: A Visual-Somatosensory fMRI Study with a Double-Flash Illusion , 2012, The Journal of Neuroscience.

[31]  E. Halgren,et al.  Cognitive response profile of the human fusiform face area as determined by MEG. , 2000, Cerebral cortex.

[32]  D. Bavelier,et al.  Cross-modal plasticity: where and how? , 2002, Nature Reviews Neuroscience.

[33]  R. Zatorre,et al.  Voice-selective areas in human auditory cortex , 2000, Nature.

[34]  Francesco Pavani,et al.  Visual Abilities in Individuals with Profound Deafness A Critical Review , 2012 .

[35]  Bruno Rossion,et al.  Temporal frequency tuning of cortical face-sensitive areas for individual face perception , 2014, NeuroImage.

[36]  K. Grill-Spector,et al.  fMR-adaptation: a tool for studying the functional properties of human cortical neurons. , 2001, Acta psychologica.

[37]  T. Hendler,et al.  Visuo-haptic object-related activation in the ventral visual pathway , 2001, Nature Neuroscience.

[38]  O. Collignon,et al.  Functional selectivity in sensory-deprived cortices. , 2011, Journal of neurophysiology.

[39]  Benton Al,et al.  Impairment in facial recognition in patients with cerebral disease. , 1968 .

[40]  N. Weisz,et al.  Revisiting the adaptive and maladaptive effects of crossmodal plasticity , 2014, Neuroscience.

[41]  Andreas Kleinschmidt,et al.  Interaction of Face and Voice Areas during Speaker Recognition , 2005, Journal of Cognitive Neuroscience.

[42]  A. Benton,et al.  Impairment in facial recognition in patients with cerebral disease. , 1968, Transactions of the American Neurological Association.

[43]  Greg O. Horne,et al.  Controlling low-level image properties: The SHINE toolbox , 2010, Behavior research methods.

[44]  P. Ackerman,et al.  Cognitive, perceptual-speed, and psychomotor determinants of individual differences during skill acquisition. , 2000, Journal of experimental psychology. Applied.

[45]  Cheryl M. Capek,et al.  The signing brain: the neurobiology of sign language , 2008, Trends in Cognitive Sciences.

[46]  J. Raven,et al.  Manual for Raven's progressive matrices and vocabulary scales , 1962 .

[47]  Karen Emmorey,et al.  Enhanced Facial Diserimination: Effects of Experience With American Sign Language , 1997 .

[48]  R. Oostenveld,et al.  Nonparametric statistical testing of EEG- and MEG-data , 2007, Journal of Neuroscience Methods.

[49]  Robert Oostenveld,et al.  FieldTrip: Open Source Software for Advanced Analysis of MEG, EEG, and Invasive Electrophysiological Data , 2010, Comput. Intell. Neurosci..

[50]  Paul Arnold,et al.  Memory for Faces and Objects by Deaf and Hearing Signers and Hearing Nonsigners , 1998 .

[51]  Alfred Anwander,et al.  Direct Structural Connections between Voice- and Face-Recognition Areas , 2011, The Journal of Neuroscience.

[52]  Joost X. Maier,et al.  Multisensory Integration of Dynamic Faces and Voices in Rhesus Monkey Auditory Cortex , 2005 .

[53]  U. Bellugi,et al.  Enhanced facial discrimination: effects of experience with American sign language. , 1997, Journal of deaf studies and deaf education.

[54]  R. Zatorre,et al.  Adaptation to speaker's voice in right anterior temporal lobe , 2003, Neuroreport.

[55]  W. Sommer,et al.  Structural invariance and age-related performance differences in face cognition. , 2010, Psychology and aging.