Multisensory Bayesian Inference Depends on Synapse Maturation during Training: Theoretical Analysis and Neural Modeling Implementation

Recent theoretical and experimental studies suggest that in multisensory conditions, the brain performs a near-optimal Bayesian estimate of external events, giving more weight to the more reliable stimuli. However, the neural mechanisms responsible for this behavior, and its progressive maturation in a multisensory environment, are still insufficiently understood. The aim of this letter is to analyze this problem with a neural network model of audiovisual integration, based on probabilistic population coding—the idea that a population of neurons can encode probability functions to perform Bayesian inference. The model consists of two chains of unisensory neurons (auditory and visual) topologically organized. They receive the corresponding input through a plastic receptive field and reciprocally exchange plastic cross-modal synapses, which encode the spatial co-occurrence of visual-auditory inputs. A third chain of multisensory neurons performs a simple sum of auditory and visual excitations. The work includes a theoretical part and a computer simulation study. We show how a simple rule for synapse learning (consisting of Hebbian reinforcement and a decay term) can be used during training to shrink the receptive fields and encode the unisensory likelihood functions. Hence, after training, each unisensory area realizes a maximum likelihood estimate of stimulus position (auditory or visual). In cross-modal conditions, the same learning rule can encode information on prior probability into the cross-modal synapses. Computer simulations confirm the theoretical results and show that the proposed network can realize a maximum likelihood estimate of auditory (or visual) positions in unimodal conditions and a Bayesian estimate, with moderate deviations from optimality, in cross-modal conditions. Furthermore, the model explains the ventriloquism illusion and, looking at the activity in the multimodal neurons, explains the automatic reweighting of auditory and visual inputs on a trial-by-trial basis, according to the reliability of the individual cues.

[1]  Yong-Di Zhou,et al.  Somatosensory cell response to an auditory cue in a haptic memory task , 2004, Behavioural Brain Research.

[2]  Thomas D. Mrsic-Flogel,et al.  Experience-Dependent Specialization of Receptive Field Surround for Selective Coding of Natural Scenes , 2014, Neuron.

[3]  R. Zemel,et al.  Inference and computation with population codes. , 2003, Annual review of neuroscience.

[4]  Mauro Ursino,et al.  Neurocomputational approaches to modelling multisensory integration in the brain: A review , 2014, Neural Networks.

[5]  A. Treves Mean-field analysis of neuronal spike dynamics , 1993 .

[6]  Wei Ji Ma,et al.  Bayesian inference with probabilistic population codes , 2006, Nature Neuroscience.

[7]  Teuvo Kohonen,et al.  Self-Organizing Maps , 2010 .

[8]  M. Wallace,et al.  Unifying multisensory signals across time and space , 2004, Experimental Brain Research.

[9]  Mauro Ursino,et al.  A Neural Network Model of Ventriloquism Effect and Aftereffect , 2012, PloS one.

[10]  John J. Foxe,et al.  Multisensory contributions to low-level, ‘unisensory’ processing , 2005, Current Opinion in Neurobiology.

[11]  David R. Wozny,et al.  Human trimodal perception follows optimal statistical inference. , 2008, Journal of vision.

[12]  Fanny Cazettes,et al.  Cue Reliability Represented in the Shape of Tuning Curves in the Owl's Sound Localization System , 2016, The Journal of Neuroscience.

[13]  FRANK MORRELL,et al.  Visual System's View of Acoustic Space , 1972, Nature.

[14]  Masakazu Konishi,et al.  Effects of Interaural Decorrelation on Neural and Behavioral Detection of Spatial Cues , 1998, Neuron.

[15]  A. Pouget,et al.  Probabilistic brains: knowns and unknowns , 2013, Nature Neuroscience.

[16]  Konrad Paul Kording,et al.  Causal Inference in Multisensory Perception , 2007, PloS one.

[17]  C. Schroeder,et al.  Neuronal mechanisms, response dynamics and perceptual functions of multisensory interactions in auditory cortex , 2009, Hearing Research.

[18]  R. Jacobs,et al.  Optimal integration of texture and motion cues to depth , 1999, Vision Research.

[19]  G. Recanzone,et al.  The biological basis of audition. , 2008, Annual review of psychology.

[20]  H. Kennedy,et al.  Anatomical Evidence of Multimodal Integration in Primate Striate Cortex , 2002, The Journal of Neuroscience.

[21]  G. Recanzone,et al.  Frequency and intensity response properties of single neurons in the auditory cortex of the behaving macaque monkey. , 2000, Journal of neurophysiology.

[22]  M T Wallace,et al.  Development of Multisensory Neurons and Multisensory Integration in Cat Superior Colliculus , 1997, The Journal of Neuroscience.

[23]  Kazuyuki Aihara,et al.  Bayesian Inference Explains Perception of Unity and Ventriloquism Aftereffect: Identification of Common Sources of Audiovisual Stimuli , 2007, Neural Computation.

[24]  Albert Jin Chung,et al.  Perception of Body Ownership Is Driven by Bayesian Sensory Inference , 2015, PloS one.

[25]  Robert A Jacobs,et al.  Bayesian integration of visual and auditory signals for spatial localization. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[26]  Christopher R Fetsch,et al.  Neural correlates of reliability-based cue weighting during multisensory integration , 2011, Nature Neuroscience.

[27]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[28]  John J. Foxe,et al.  Hebbian Learning Mechanisms Help Explain the Maturation of Multisensory Speech Integration in Children with Autism Spectrum Disorder (ASD) and with Typical Development (TD): a Neurocomputational Analysis , 2015, EAPCogSci.

[29]  M. Wallace,et al.  Visual Localization Ability Influences Cross-Modal Bias , 2003, Journal of Cognitive Neuroscience.

[30]  Kathleen S Rockland,et al.  Multisensory convergence in calcarine visual areas in macaque monkey. , 2003, International journal of psychophysiology : official journal of the International Organization of Psychophysiology.

[31]  J. Fuster,et al.  Visuo-tactile cross-modal associations in cortical somatosensory cells. , 2000, Proceedings of the National Academy of Sciences of the United States of America.

[32]  Wei Ji Ma,et al.  Efficient Probabilistic Inference in Generic Neural Networks Trained with Non-Probabilistic Feedback , 2018 .

[33]  Ulrik R Beierholm,et al.  Sound-induced flash illusion as an optimal percept , 2005, Neuroreport.

[34]  Konrad Paul Kording,et al.  Over my fake body: body ownership illusions for studying the multisensory basis of own-body perception , 2015, Front. Hum. Neurosci..

[35]  A. Ghazanfar,et al.  Is neocortex essentially multisensory? , 2006, Trends in Cognitive Sciences.

[36]  J. Driver,et al.  Multisensory Interplay Reveals Crossmodal Influences on ‘Sensory-Specific’ Brain Regions, Neural Responses, and Judgments , 2008, Neuron.

[37]  G. Recanzone Interactions of auditory and visual stimuli in space and time , 2009, Hearing Research.

[38]  Mauro Ursino,et al.  A Neural Network Model Can Explain Ventriloquism Aftereffect and Its Generalization across Sound Frequencies , 2013, BioMed research international.

[39]  G. DeAngelis,et al.  A Normalization Model of Multisensory Integration , 2011, Nature Neuroscience.

[40]  Robert C. Froemke,et al.  Development of auditory cortical synaptic receptive fields , 2011, Neuroscience & Biobehavioral Reviews.

[41]  D. Burr,et al.  The Ventriloquist Effect Results from Near-Optimal Bimodal Integration , 2004, Current Biology.

[42]  Nadia Bolognini,et al.  A neurocomputational analysis of the sound-induced flash illusion , 2014, NeuroImage.

[43]  G. DeAngelis,et al.  Multisensory Integration in Macaque Visual Cortex Depends on Cue Reliability , 2008, Neuron.

[44]  Anders Krogh,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[45]  Mikko Sams,et al.  Factors influencing audiovisual fission and fusion illusions. , 2004, Brain research. Cognitive brain research.

[46]  W. Ma,et al.  Towards a neural implementation of causal inference in cue combination. , 2013, Multisensory research.

[47]  S. Jbabdi,et al.  How can a Bayesian approach inform neuroscience? , 2012, The European journal of neuroscience.

[48]  James M. Hillis,et al.  Slant from texture and disparity cues: optimal cue combination. , 2004, Journal of vision.

[49]  Ulrik R. Beierholm,et al.  Causal inference in perception , 2010, Trends in Cognitive Sciences.

[50]  H. McGurk,et al.  Hearing lips and seeing voices , 1976, Nature.

[51]  Brian J. Fischer,et al.  Owl's behavior and neural representation predicted by Bayesian inference , 2011, Nature Neuroscience.

[52]  S. Shimojo,et al.  Illusions: What you see is what you hear , 2000, Nature.

[53]  P. Bertelson,et al.  Cross-modal bias and perceptual fusion with auditory-visual spatial discordance , 1981, Perception & psychophysics.

[54]  Thomas J. Anastasio,et al.  Modeling Cross-Modal Enhancement and Modality-Specific Suppression in Multisensory Neurons , 2003, Neural Computation.