Multisensory perception reflects individual differences in processing temporal correlations

Sensory signals originating from a single event, such as audiovisual speech, are temporally correlated. Correlated signals are known to facilitate multisensory integration and binding. We sought to further elucidate the nature of this relationship, hypothesizing that multisensory perception will vary with the strength of audiovisual correlation. Human participants detected near-threshold amplitude modulations in auditory and/or visual stimuli. During audiovisual trials, the frequency and phase of auditory modulations were varied, producing signals with a range of correlations. After accounting for individual differences which likely reflect relative unisensory temporal characteristics in participants, we found that multisensory perception varied linearly with strength of correlation. Diffusion modelling confirmed this and revealed that stimulus correlation is supplied to the decisional system as sensory evidence. These data implicate correlation as an important cue in audiovisual feature integration and binding and suggest correlational strength as an important factor for flexibility in these processes.

[1]  Daniel Senkowski,et al.  Good times for multisensory integration: Effects of the precision of temporal synchrony as revealed by gamma-band oscillations , 2007, Neuropsychologia.

[2]  James V. Stone,et al.  When is now? Perception of simultaneity , 2001, Proceedings of the Royal Society of London. Series B: Biological Sciences.

[3]  D. M. Green,et al.  Signal detection theory and psychophysics , 1966 .

[4]  J. Obleser,et al.  Frequency modulation entrains slow neural oscillations and optimizes human listening behavior , 2012, Proceedings of the National Academy of Sciences.

[5]  G. H. Bishop,et al.  CYCLIC CHANGES IN EXCITABILITY OF THE OPTIC PATHWAY OF THE RABBIT , 1932 .

[6]  C. Koch,et al.  The Neural Correlates of Consciousness , 2008, Annals of the New York Academy of Sciences.

[7]  A. Voss,et al.  Interpreting the parameters of the diffusion model: An empirical validation , 2004, Memory & cognition.

[8]  M HERSHENSON,et al.  Reaction time as a measure of intersensory facilitation. , 1962, Journal of experimental psychology.

[9]  Ladan Shams,et al.  Biases in Visual, Auditory, and Audiovisual Perception of Space , 2015, PLoS Comput. Biol..

[10]  W R Thurlow,et al.  Effects of degree of visual association and angle of displacement on the "ventriloquism" effect. , 1973, Perceptual and motor skills.

[11]  J. Gold,et al.  The neural basis of decision making. , 2007, Annual review of neuroscience.

[12]  Andreas K. Engel,et al.  Oscillatory Synchronization in Large-Scale Cortical Networks Predicts Perception , 2011, Neuron.

[13]  Marc O. Ernst,et al.  Correlation detection as a general mechanism for multisensory integration , 2016, Nature Communications.

[14]  Randolph Blake,et al.  The role of temporal structure in human vision. , 2005, Behavioral and cognitive neuroscience reviews.

[15]  John J. Foxe,et al.  Recalibration of the Multisensory Temporal Window of Integration Results from Changing Task Demands , 2013, PloS one.

[16]  F. Leite A comparison of two diffusion process models in accounting for payoff and stimulus frequency manipulations , 2012, Attention, Perception, & Psychophysics.

[17]  W. H. Sumby,et al.  Visual contribution to speech intelligibility in noise , 1954 .

[18]  Isabelle Peretz,et al.  Steady-state evoked potentials as an index of multisensory temporal binding , 2012, NeuroImage.

[19]  J Driver,et al.  Crossmodal Spatial Influences of Touch on Extrastriate Visual Areas Take Current Gaze Direction into Account , 2002, Neuron.

[20]  Wei Ji Ma,et al.  Causal inference of asynchronous audiovisual speech , 2013, Front. Psychol..

[21]  Allison M. McKendrick,et al.  Response times across the visual field: Empirical observations and application to threshold determination , 2014, Vision Research.

[22]  Adrian K. C. Lee,et al.  Auditory selective attention is enhanced by a task-irrelevant temporally coherent visual stimulus in human listeners , 2015, eLife.

[23]  D H Brainard,et al.  The Psychophysics Toolbox. , 1997, Spatial vision.

[24]  Hans Colonius,et al.  On quantifying multisensory interaction effects in reaction time and detection rate , 2011, Psychological research.

[25]  B. Stein The new handbook of multisensory processes , 2012 .

[26]  B. Stein,et al.  Spatial factors determine the activity of multisensory neurons in cat superior colliculus , 1986, Brain Research.

[27]  W Singer,et al.  Visual feature integration and the temporal correlation hypothesis. , 1995, Annual review of neuroscience.

[28]  M. Wallace,et al.  Representation and integration of multiple sensory inputs in primate superior colliculus. , 1996, Journal of neurophysiology.

[29]  C. Schroeder,et al.  Low-frequency neuronal oscillations as instruments of sensory selection , 2009, Trends in Neurosciences.

[30]  John J. Foxe,et al.  Crossmodal binding through neural coherence: implications for multisensory processing , 2008, Trends in Neurosciences.

[31]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[32]  Philip L. Smith,et al.  The accumulator model of two-choice discrimination , 1988 .

[33]  D. Burr,et al.  The Ventriloquist Effect Results from Near-Optimal Bimodal Integration , 2004, Current Biology.

[34]  M. Hallett,et al.  Neural Correlates of Auditory–Visual Stimulus Onset Asynchrony Detection , 2001, The Journal of Neuroscience.

[35]  P. Fries A mechanism for cognitive dynamics: neuronal communication through neuronal coherence , 2005, Trends in Cognitive Sciences.

[36]  D. Alais,et al.  Rapid Recalibration to Audiovisual Asynchrony , 2013, The Journal of Neuroscience.

[37]  N. Bolognini,et al.  Enhancement of visual perception by crossmodal visuo-auditory interaction , 2002, Experimental Brain Research.

[38]  Aaron R. Nidiffer,et al.  Single Trial Plasticity in Evidence Accumulation Underlies Rapid Recalibration to Asynchronous Audiovisual Speech , 2018, Scientific Reports.

[39]  Michael Schutz,et al.  The unity assumption facilitates cross-modal binding of musical, non-speech stimuli: The role of spectral and amplitude envelope cues , 2016, Attention, Perception, & Psychophysics.

[40]  A. Diederich,et al.  The time window of multisensory integration: relating reaction times and judgments of temporal order. , 2015, Psychological review.

[41]  E Macaluso,et al.  Spatial and temporal factors during processing of audiovisual speech: a PET study , 2004, NeuroImage.

[42]  Roger Ratcliff,et al.  The Diffusion Decision Model: Theory and Data for Two-Choice Decision Tasks , 2008, Neural Computation.

[43]  Jennifer K Bizley,et al.  Where are multisensory signals combined for perceptual decision-making? , 2016, Current Opinion in Neurobiology.

[44]  S. Shamma,et al.  Temporal Coherence in the Perceptual Organization and Cortical Representation of Auditory Scenes , 2009, Neuron.

[45]  S. Nishida,et al.  Recalibration of audiovisual simultaneity , 2004, Nature Neuroscience.

[46]  A. Diederich,et al.  Multisensory Interaction in Saccadic Reaction Time: A Time-Window-of-Integration Model , 2004, Journal of Cognitive Neuroscience.

[47]  Donald Laming,et al.  Information theory of choice-reaction times , 1968 .

[48]  Adele Diederich,et al.  Simple matrix methods for analyzing diffusion models of choice probability, choice response time, and simple response time , 2003 .

[49]  Adrian K. C. Lee,et al.  Defining Auditory-Visual Objects: Behavioral Tests and Physiological Mechanisms , 2016, Trends in Neurosciences.

[50]  Jeffrey C. Lagarias,et al.  Convergence Properties of the Nelder-Mead Simplex Method in Low Dimensions , 1998, SIAM J. Optim..

[51]  Jonathan D. Cohen,et al.  The physics of optimal decision making: a formal analysis of models of performance in two-alternative forced-choice tasks. , 2006, Psychological review.

[52]  N. Bolognini,et al.  “Acoustical vision” of below threshold stimuli: interaction among spatially converging audiovisual inputs , 2004, Experimental Brain Research.

[53]  A. Diederich,et al.  Modeling the effects of payoff on response bias in a perceptual discrimination task: Bound-change, drift-rate-change, or two-stage-processing hypothesis , 2006, Perception & psychophysics.

[54]  P. Gribble,et al.  Temporal constraints on the McGurk effect , 1996, Perception & psychophysics.

[55]  M. Frens,et al.  Spatial and temporal factors determine auditory-visual interactions in human saccadic eye movements , 1995, Perception & psychophysics.

[56]  B. Stein,et al.  Determinants of multisensory integration in superior colliculus neurons. I. Temporal factors , 1987, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[57]  P. Schyns,et al.  Entrainment of Perceptually Relevant Brain Oscillations by Non-Invasive Rhythmic Stimulation of the Human Brain , 2011, Front. Psychology.

[58]  C. Spence,et al.  Crossmodal binding: Evaluating the “unity assumption” using audiovisual speech stimuli , 2007, Perception & psychophysics.

[59]  J. Obleser,et al.  Entrained neural oscillations in multiple frequency bands comodulate behavior , 2014, Proceedings of the National Academy of Sciences.

[60]  Q. Summerfield,et al.  Intermodal timing relations and audio-visual speech recognition by normal-hearing adults. , 1985, The Journal of the Acoustical Society of America.

[61]  Alexandre Gramfort,et al.  Encoding of event timing in the phase of neural oscillations , 2014, NeuroImage.

[62]  Charles Spence,et al.  Cross-correlation between auditory and visual signals promotes multisensory integration. , 2013, Multisensory research.

[63]  Adrian K. C. Lee,et al.  Integration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding , 2017, Neuron.

[64]  Denis G. Pelli,et al.  ECVP '07 Abstracts , 2007, Perception.

[65]  N. P. Erber Interaction of audition and vision in the recognition of oral speech stimuli. , 1969, Journal of speech and hearing research.

[66]  W. Singer,et al.  Temporal binding and the neural correlates of sensory awareness , 2001, Trends in Cognitive Sciences.

[67]  Philip L. Smith,et al.  A comparison of sequential sampling models for two-choice reaction time. , 2004, Psychological review.

[68]  N. F. Dixon,et al.  The Detection of Auditory Visual Desynchrony , 1980, Perception.

[69]  Asif A. Ghazanfar,et al.  The Natural Statistics of Audiovisual Speech , 2009, PLoS Comput. Biol..

[70]  Philipp Berens,et al.  CircStat: AMATLABToolbox for Circular Statistics , 2009, Journal of Statistical Software.

[71]  P F Seitz,et al.  The use of visible speech cues for improving auditory detection of spoken sentences. , 2000, The Journal of the Acoustical Society of America.

[72]  M. Ernst,et al.  When Correlation Implies Causation in Multisensory Integration , 2012, Current Biology.

[73]  Konrad Paul Kording,et al.  Causal Inference in Multisensory Perception , 2007, PloS one.

[74]  M. A. Frens,et al.  A quantitative study of auditory-evoked saccadic eye movements in two dimensions , 2004, Experimental Brain Research.

[75]  Thomas U. Otto,et al.  Noise and Correlations in Parallel Perceptual Decision Making , 2011, Current Biology.