What you see depends on what you hear: temporal averaging and crossmodal integration

In our multisensory world, we often rely more on auditory information than on visual input for temporal processing. One typical demonstration of this is that the rate of auditory flutter assimilates the rate of concurrent visual flicker. To date, however, this auditory dominance effect has largely been studied using regular auditory rhythms. It thus remains unclear whether irregular rhythms would have a similar impact on visual temporal processing, what information is extracted from the auditory sequence that comes to influence visual timing, and how the auditory and visual temporal rates are integrated together in quantitative terms. We investigated these questions by assessing, and modeling, the influence of a task-irrelevant auditory sequence on the type of “Ternus apparent motion”: group motion versus element motion. The type of motion seen critically depends on the time interval between the two Ternus display frames. We found that an irrelevant auditory sequence preceding the Ternus display modulates the visual interval, making observers perceive either more group motion or more element motion. This biasing effect manifests whether the auditory sequence is regular or irregular, and it is based on a summary statistic extracted from the sequential intervals: their geometric mean. However, the audiovisual interaction depends on the discrepancy between the mean auditory and visual intervals: if it becomes too large, no interaction occurs—which can be quantitatively described by a partial Bayesian integration model. Overall, our findings reveal a cross-modal perceptual averaging principle that may underlie complex audiovisual interactions in many everyday dynamic situations.

[1]  Michael N. Shadlen,et al.  Temporal context calibrates interval timing , 2010, Nature Neuroscience.

[2]  Melissa E. Libertus,et al.  Comment on "Log or Linear? Distinct Intuitions of the Number Scale in Western and Amazonian Indigene Cultures" , 2009, Science.

[3]  Laura A. Carlson,et al.  When What You Hear Influences When You See , 2013, Psychological science.

[4]  Neil W Roach,et al.  Generalization of prior information for rapid Bayesian time estimation , 2016, Proceedings of the National Academy of Sciences.

[5]  Dean V Buonomano,et al.  Neurocomputational models of interval and pattern timing , 2016, Current Opinion in Behavioral Sciences.

[6]  Robert B. Welch,et al.  Contributions of audition and vision to temporal rate perception , 1986, Perception & psychophysics.

[7]  Warren H. Meck,et al.  Bayesian optimization of time perception , 2013, Trends in Cognitive Sciences.

[8]  M. Ernst,et al.  When Correlation Implies Causation in Multisensory Integration , 2012, Current Biology.

[9]  Felix Wichmann,et al.  The psychometric function: I , 2001 .

[10]  Stanislas Dehaene,et al.  The neural basis of the Weber–Fechner law: a logarithmic mental number line , 2003, Trends in Cognitive Sciences.

[11]  Konrad Paul Kording,et al.  Causal Inference in Multisensory Perception , 2007, PloS one.

[12]  Jüri Allik,et al.  Obligatory averaging in mean size perception , 2014, Vision Research.

[13]  Michael A. Cohen,et al.  What is the Bandwidth of Perceptual Experience? , 2016, Trends in Cognitive Sciences.

[14]  Brigitte Röder,et al.  Audiotactile integration is reduced in congenital blindness in a spatial ventriloquism task , 2012, Neuropsychologia.

[15]  Ryota Kanai,et al.  Modality-independent role of the primary auditory cortex in time estimation , 2011, Experimental Brain Research.

[16]  Joan López-Moliner,et al.  quickpsy: An R Package to Fit Psychometric Functions for Multiple Groups , 2016, R J..

[17]  Eero P. Simoncelli,et al.  Article Sound Texture Perception via Statistics of the Auditory Periphery: Evidence from Sound Synthesis , 2022 .

[18]  D. Ariely Seeing Sets: Representation by Statistical Properties , 2001, Psychological science.

[19]  G. Knox Investigations of Flicker and Fusion: IV. The Effect of Auditory Flicker on the Pronouncedness of Visual Flicker , 1945 .

[20]  R. Marois,et al.  Capacity limits of information processing in the brain , 2005, Trends in Cognitive Sciences.

[21]  A. K. Myers,et al.  Matching the rate of concurrent tone bursts and light flashes as a function of flash surround luminance , 1981, Perception & psychophysics.

[22]  R. Church,et al.  Application of scalar timing theory to individual trials. , 1994, Journal of experimental psychology. Animal behavior processes.

[23]  G. Recanzone,et al.  Temporal and spatial dependency of the ventriloquism effect , 2001, Neuroreport.

[24]  Marc O. Ernst,et al.  Correlation detection as a general mechanism for multisensory integration , 2016, Nature Communications.

[25]  D. Whitaker,et al.  Recalibration of perceived time across sensory modalities , 2008, Experimental Brain Research.

[26]  C. Bundesen,et al.  A neural theory of visual attention: bridging cognition and neurophysiology. , 2005, Psychological review.

[27]  Randolph Blake,et al.  Hearing What the Eyes See , 2005, Psychological science.

[28]  D H Brainard,et al.  The Psychophysics Toolbox. , 1997, Spatial vision.

[29]  Nelson Cowan,et al.  Metatheory of storage capacity limits , 2001, Behavioral and Brain Sciences.

[30]  E. Miller,et al.  Coding of Cognitive Magnitude Compressed Scaling of Numerical Information in the Primate Prefrontal Cortex , 2003, Neuron.

[31]  Zhuanghua Shi,et al.  Auditory temporal modulation of the visual Ternus effect: the influence of time interval , 2010, Experimental Brain Research.

[32]  M. Concetta Morrone,et al.  Contextual effects in interval-duration judgements in vision, audition and touch , 2013, Experimental Brain Research.

[33]  M. Jones,et al.  Temporal Aspects of Stimulus-Driven Attending in Dynamic Arrays , 2002, Psychological science.

[34]  S Buus Temporal integration and multiple looks, revisited: weights as a function of time. , 1999, The Journal of the Acoustical Society of America.

[35]  D. Burr,et al.  Optimal Encoding of Interval Timing in Expert Percussionists , 2011, The Journal of Neuroscience.

[36]  D. Melcher,et al.  The Role of Oscillatory Phase in Determining the Temporal Organization of Perception: Evidence from Sensory Entrainment , 2017, The Journal of Neuroscience.

[37]  Neil W. Roach,et al.  Resolving multisensory conflict: a strategy for balancing the costs and benefits of audio-visual integration , 2006, Proceedings of the Royal Society B: Biological Sciences.

[38]  Yoshitaka Nakajima,et al.  Time-Shrinking: The Process of Unilateral Temporal Assimilation , 2004, Perception.

[39]  J. Gibbon,et al.  Human bisection at the geometric mean , 1991 .

[40]  F A Wichmann,et al.  Ning for Helpful Comments and Suggestions. This Paper Benefited Con- Siderably from Conscientious Peer Review, and We Thank Our Reviewers the Psychometric Function: I. Fitting, Sampling, and Goodness of Fit , 2001 .

[41]  J. Gibbon Scalar expectancy theory and Weber's law in animal timing. , 1977 .

[42]  G. Alvarez Representing multiple objects as an ensemble enhances visual cognition , 2011, Trends in Cognitive Sciences.

[43]  G. Recanzone Auditory influences on visual temporal rate perception. , 2003, Journal of neurophysiology.

[44]  Yoshitaka Nakajima,et al.  Time-shrinking: A discontinuity in the perception of auditory temporal patterns , 1992, Perception & psychophysics.

[45]  G. Karmos,et al.  Entrainment of Neuronal Oscillations as a Mechanism of Attentional Selection , 2008, Science.

[46]  G. Recanzone Interactions of auditory and visual stimuli in space and time , 2009, Hearing Research.

[47]  E. Large,et al.  The dynamics of attending: How people track time-varying events. , 1999 .

[48]  Lance J Rips,et al.  How many is a zillion? Sources of number distortion. , 2013, Journal of experimental psychology. Learning, memory, and cognition.

[49]  T SHIPLEY,et al.  Auditory Flutter-Driving of Visual Flicker , 1964, Science.

[50]  T. Bayne,et al.  Ensemble Coding and Two Conceptions of Perceptual Sparsity , 2016, Trends in Cognitive Sciences.

[51]  Andreas Nieder,et al.  A parieto-frontal network for visual numerical information in the monkey. , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[52]  G H MOWBRAY,et al.  On discriminating the rate of visual flicker and auditory flutter. , 1959, The American journal of psychology.

[53]  F. Petzschner,et al.  A Bayesian perspective on magnitude estimation , 2015, Trends in Cognitive Sciences.

[54]  A. Chetverikov,et al.  Building ensemble representations: How the shape of preceding distractor distributions affects visual search , 2016, Cognition.

[55]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[56]  D. Burr,et al.  Predictive coding of multisensory timing , 2016, Current Opinion in Behavioral Sciences.

[57]  J. Vroomen,et al.  Intersensory binding across space and time: A tutorial review , 2013, Attention, Perception, & Psychophysics.

[58]  David Whitaker,et al.  Audiovisual time perception is spatially specific , 2012, Experimental Brain Research.

[59]  Vincent Walsh A theory of magnitude: common cortical metrics of time, space and quantity , 2003, Trends in Cognitive Sciences.