Cue Combination Within a Bayesian Framework

To interact effectively with the world, the brain must optimize its perception of the objects and events in the environment, many of which are signaled by more than one sense. Optimal perception requires the brain to integrate redundant cues from the different senses as efficiently as possible. One effective model of cue combination is maximum likelihood estimation (MLE), a Bayesian model that deals with the fundamental uncertainty and noise associated with sensory signals and provides a statistically optimal way to integrate them. MLE achieves this through a weighted linear sum of two or more cues in which each cue is weighted inversely to its variance or “uncertainty.” This produces an integrated sensory estimate with minimal uncertainty and thus maximized perceptual precision. Many studies show that adults integrate redundant sensory information consistent with MLE predictions. When the MLE model is tested in school-aged children, it is found that predictions for multisensory integration are confirmed in older children (>10 years) but not in younger children. Younger children show unisensory dominance and do not exploit the statistical benefits of multisensory integration, even when their dominant sense is far less precise than the other. This curious finding may result from each sensory system having an inherent specialization, with each specialist sense tuning the other senses, such as vision calibrating audition for space (or audition calibrating vision for time). This cross-sensory tuning would preclude useful combination of two senses until calibration is complete, after which MLE integration provides an excellent model of multisensory cue combination.

[1]  H. Bülthoff,et al.  Merging the senses into a robust percept , 2004, Trends in Cognitive Sciences.

[2]  Hamish Innes-Brown,et al.  Audiovisual integration in noise by children and adults. , 2010, Journal of experimental child psychology.

[3]  James J. Clark,et al.  Data Fusion for Sensory Information Processing Systems , 1990 .

[4]  Angela M. Brown,et al.  Visual acuity of human infants at scotopic, mesopic and photopic luminances , 1987, Vision Research.

[5]  D. Alais,et al.  Temporal ventriloquism in a purely temporal context. , 2011, Journal of experimental psychology. Human perception and performance.

[6]  Barbara Dodd,et al.  Lip reading in infants: Attention to speech presented in- and out-of-synchrony , 1979, Cognitive Psychology.

[7]  Gregor Thut,et al.  The multisensory function of the human primary visual cortex , 2016, Neuropsychologia.

[8]  Dan Farrell Objective Bayesian Analysis in Acoustics - Ning Xiang and Cameron Fackler , 2015 .

[9]  G. Recanzone,et al.  Temporal and spatial dependency of the ventriloquism effect , 2001, Neuroreport.

[10]  S. Rose Developmental changes in infants' retention of visual stimuli. , 1981, Child development.

[11]  L. Benevento,et al.  Auditory-visual interaction in single cells in the cortex of the superior temporal sulcus and the orbital frontal cortex of the macaque monkey , 1977, Experimental Neurology.

[12]  Wayne A. Hershberger,et al.  Haptic estimates of discordant visual—haptic size vary developmentally , 1999, Perception & psychophysics.

[13]  Andrew J. King,et al.  Visual influences on auditory spatial learning , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.

[14]  D. Burr,et al.  Auditory dominance over vision in the perception of interval duration , 2009, Experimental Brain Research.

[15]  D. Hubel,et al.  Receptive fields and functional architecture of monkey striate cortex , 1968, The Journal of physiology.

[16]  G. Gottlieb Development of species identification in birds : an inquiry into the prenatal determinants of perception , 1971 .

[17]  I. Howard,et al.  Human Spatial Orientation , 1966 .

[18]  D. Lewkowicz,et al.  The development of intersensory temporal perception: an epigenetic systems/limitations view. , 2000, Psychological bulletin.

[19]  G. Recanzone Auditory influences on visual temporal rate perception. , 2003, Journal of neurophysiology.

[20]  D. Burr,et al.  Predictive coding of multisensory timing , 2016, Current Opinion in Behavioral Sciences.

[21]  C. Spence,et al.  Tactile “capture” of audition , 2002, Perception & psychophysics.

[22]  D. Knill,et al.  The Bayesian brain: the role of uncertainty in neural coding and computation , 2004, Trends in Neurosciences.

[23]  T. Powell,et al.  An anatomical study of converging sensory pathways within the cerebral cortex of the monkey. , 1970, Brain : a journal of neurology.

[24]  D. Knill Learning Bayesian priors for depth perception. , 2007, Journal of vision.

[25]  B A Schneider,et al.  Gap detection in infants, children, and adults. , 1995, The Journal of the Acoustical Society of America.

[26]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[27]  G. DeAngelis,et al.  Neural correlates of multisensory cue integration in macaque MSTd , 2008, Nature Neuroscience.

[28]  T SHIPLEY,et al.  Auditory Flutter-Driving of Visual Flicker , 1964, Science.

[29]  James M. Hillis,et al.  Combining Sensory Information: Mandatory Fusion Within, but Not Between, Senses , 2002, Science.

[30]  G. Sandini,et al.  Development of Visuo-Auditory Integration in Space and Time , 2012, Front. Integr. Neurosci..

[31]  David C. Burr,et al.  Young Children Do Not Integrate Visual and Haptic Form Information , 2008, Current Biology.

[32]  G. Sandini,et al.  Poor Haptic Orientation Discrimination in Nonsighted Children May Reflect Disruption of Cross-Sensory Calibration , 2010, Current Biology.

[33]  S. Shimojo,et al.  Illusions: What you see is what you hear , 2000, Nature.

[34]  Yvette Hatwell,et al.  Motor and Cognitive Functions of the Hand in Infancy and Childhood , 1987 .

[35]  Mowbray Gh,et al.  On discriminating the rate of visual flicker and auditory flutter. , 1959 .

[36]  Marc O Ernst,et al.  The Duration of Uncertain Times: Audiovisual Information about Intervals Is Integrated in a Statistically Optimal Fashion , 2014, PloS one.

[37]  G Turkewitz,et al.  Intersensory interaction in newborns: modification of visual preferences following exposure to sound. , 1981, Child development.

[38]  D. H. Warren,et al.  Immediate perceptual response to intersensory discrepancy. , 1980, Psychological bulletin.

[39]  J. Saunders,et al.  Do humans optimally integrate stereo and texture information for judgments of surface slant? , 2003, Vision Research.

[40]  P. Bertelson,et al.  Cross-modal bias and perceptual fusion with auditory-visual spatial discordance , 1981, Perception & psychophysics.

[41]  D. Lewkowicz Infants’ responsiveness to the auditory and visual attributes of a sounding/moving stimulus , 1992, Perception & psychophysics.

[42]  James M. Hillis,et al.  Slant from texture and disparity cues: optimal cue combination. , 2004, Journal of vision.

[43]  Hermann von Helmholtz,et al.  Treatise on Physiological Optics , 1962 .

[44]  H. McGurk,et al.  Intermodal coordination in young children: Vision and touch. , 1980 .

[45]  David R. Wozny,et al.  Human trimodal perception follows optimal statistical inference. , 2008, Journal of vision.

[46]  Robert Lickliter,et al.  Infants’ perception of rhythm and tempo in unimodal and multimodal stimulation: A developmental test of the intersensory redundancy hypothesis , 2004, Cognitive, affective & behavioral neuroscience.

[47]  D. H. Warren,et al.  The role of visual-auditory “compellingness” in the ventriloquism effect: Implications for transitivity among the spatial senses , 1981, Perception & psychophysics.

[48]  G. Sandini,et al.  Impairment of auditory spatial localization in congenitally blind human subjects , 2013, Brain : a journal of neurology.

[49]  R. J. van Beers,et al.  Integration of proprioceptive and visual position-information: An experimentally supported model. , 1999, Journal of neurophysiology.

[50]  N. Logothetis,et al.  Visual modulation of neurons in auditory cortex. , 2008, Cerebral cortex.

[51]  I ROCK,et al.  Vision and Touch: An Experimentally Created Conflict between the Two Senses , 1964, Science.

[52]  A. Yuille,et al.  Object perception as Bayesian inference. , 2004, Annual review of psychology.

[53]  Michael N. Shadlen,et al.  Temporal context calibrates interval timing , 2010, Nature Neuroscience.

[54]  C. Spence,et al.  Audiotactile interactions in roughness perception , 2002, Experimental Brain Research.

[55]  A. Streri,et al.  Cross-modal recognition of shape from hand to eyes in human newborns , 2003, Somatosensory & motor research.

[56]  D. Burr,et al.  The Ventriloquist Effect Results from Near-Optimal Bimodal Integration , 2004, Current Biology.

[57]  S. Crewther,et al.  The race that precedes coactivation: development of multisensory facilitation in children. , 2009, Developmental science.

[58]  Jean-Pierre Bresciani,et al.  Vision and touch are automatically integrated for the perception of sequences of events. , 2006, Journal of vision.

[59]  Konrad Paul Kording,et al.  Bayesian integration in sensorimotor learning , 2004, Nature.

[60]  D. Knill,et al.  The brain uses adaptive internal models of scene statistics for sensorimotor estimation and planning , 2013, Proceedings of the National Academy of Sciences.

[61]  P. Mamassian,et al.  Multisensory processing in review: from physiology to behaviour. , 2010, Seeing and perceiving.

[62]  A. Pouget,et al.  Probabilistic brains: knowns and unknowns , 2013, Nature Neuroscience.

[63]  J. Werker,et al.  Infants' ability to match dynamic phonetic and gender information in the face and voice. , 2002, Journal of experimental child psychology.

[64]  Zoubin Ghahramani,et al.  Modular decomposition in visuomotor learning , 1997, Nature.

[65]  P. Bertelson,et al.  The ventriloquist effect does not depend on the direction of automatic visual attention , 2001, Perception & psychophysics.

[66]  G. DeAngelis,et al.  Multisensory integration: psychophysics, neurophysiology, and computation , 2009, Current Opinion in Neurobiology.

[67]  Virginie van Wassenhove,et al.  Auditory-visual fusion in speech perception in children with cochlear implants. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[68]  D. H. Warren,et al.  Sensory conflict in judgments of spatial direction , 1969 .

[69]  M. Landy,et al.  Ideal-Observer Models of Cue Integration , 2012 .

[70]  Pete R. Jones,et al.  Development of Cue Integration in Human Navigation , 2008, Current Biology.

[71]  Barry Stein,et al.  A Bayesian model unifies multisensory spatial localization with the physiological properties of the superior colliculus , 2007, Experimental Brain Research.

[72]  Marc O Ernst,et al.  Visual-haptic cue weighting is independent of modality-specific attention. , 2008, Journal of vision.

[73]  D. Burr,et al.  Optimal Encoding of Interval Timing in Expert Percussionists , 2011, The Journal of Neuroscience.

[74]  L. W. Olsho Infant frequency discrimination , 1984 .

[75]  Rolf Ulrich,et al.  Crossmodal temporal discrimination: Assessing the predictions of a general pacemaker-counter model , 2006, Perception & psychophysics.

[76]  N. B. Spetner,et al.  Pure‐tone sensitivity of human infants , 1986 .

[77]  M. Nardini,et al.  Fusion of visual cues is not mandatory in children , 2010, Proceedings of the National Academy of Sciences.

[78]  T. Paus Mapping brain development and aggression. , 2005, The Canadian child and adolescent psychiatry review = La revue canadienne de psychiatrie de l'enfant et de l'adolescent.

[79]  Monica Gori,et al.  Perceived duration of Visual and Tactile Stimuli Depends on Perceived Speed , 2011, Front. Integr. Neurosci..