Cross-correlation between auditory and visual signals promotes multisensory integration.

Humans are equipped with multiple sensory channels that provide both redundant and complementary information about the objects and events in the world around them. A primary challenge for the brain is therefore to solve the 'correspondence problem', that is, to bind those signals that likely originate from the same environmental source, while keeping separate those unisensory inputs that likely belong to different objects/events. Whether multiple signals have a common origin or not must, however, be inferred from the signals themselves through a causal inference process. Recent studies have demonstrated that cross-correlation, that is, the similarity in temporal structure between unimodal signals, represents a powerful cue for solving the correspondence problem in humans. Here we provide further evidence for the role of the temporal correlation between auditory and visual signals in multisensory integration. Capitalizing on the well-known fact that sensitivity to crossmodal conflict is inversely related to the strength of coupling between the signals, we measured sensitivity to crossmodal spatial conflicts as a function of the cross-correlation between the temporal structures of the audiovisual signals. Observers' performance was systematically modulated by the cross-correlation, with lower sensitivity to crossmodal conflict being measured for correlated as compared to uncorrelated audiovisual signals. These results therefore provide support for the claim that cross-correlation promotes multisensory integration. A Bayesian framework is proposed to interpret the present results, whereby stimulus correlation is represented on the prior distribution of expected crossmodal co-occurrence.

[1]  A. Watson,et al.  Quest: A Bayesian adaptive psychometric method , 1983, Perception & psychophysics.

[2]  M. Ernst Learning to integrate arbitrary signals from vision and touch. , 2007, Journal of vision.

[3]  F A Wichmann,et al.  Ning for Helpful Comments and Suggestions. This Paper Benefited Con- Siderably from Conscientious Peer Review, and We Thank Our Reviewers the Psychometric Function: I. Fitting, Sampling, and Goodness of Fit , 2001 .

[4]  Charles Spence,et al.  ‘When Birds of a Feather Flock Together’: Synesthetic Correspondences Modulate Audiovisual Integration in Non-Synesthetes , 2009, PloS one.

[5]  Max Wertheimer,et al.  Untersuchungen zur Lehre von der Gestalt , .

[6]  David Raposo,et al.  Multisensory Decision-Making in Rats and Humans , 2012, The Journal of Neuroscience.

[7]  D. H. Warren,et al.  Immediate perceptual response to intersensory discrepancy. , 1980, Psychological bulletin.

[8]  James M. Hillis,et al.  Combining Sensory Information: Mandatory Fusion Within, but Not Between, Senses , 2002, Science.

[9]  C. Spence,et al.  Crossmodal binding: Evaluating the “unity assumption” using audiovisual speech stimuli , 2007, Perception & psychophysics.

[10]  S. Gepshtein,et al.  The combination of vision and touch depends on spatial proximity. , 2005, Journal of vision.

[11]  R Blake,et al.  Visual form created solely from temporal structure. , 1999, Science.

[12]  M. Ernst,et al.  When correlation implies causation in multisensory integration , 2012 .

[13]  Denis G. Pelli,et al.  ECVP '07 Abstracts , 2007, Perception.

[14]  Jean-Pierre Bresciani,et al.  Vision and touch are automatically integrated for the perception of sequences of events. , 2006, Journal of vision.

[15]  M. Ernst,et al.  Feeling what you hear: auditory signals can modulate tactile tap perception , 2005, Experimental Brain Research.

[16]  Waka Fujisaki,et al.  Temporal frequency characteristics of synchrony–asynchrony discrimination of audio-visual signals , 2005, Experimental Brain Research.

[17]  Konrad Paul Kording,et al.  Sensory Cue Integration , 2011 .

[18]  Konrad Paul Kording,et al.  Causal Inference in Multisensory Perception , 2007, PloS one.

[19]  H. Bülthoff,et al.  Merging the senses into a robust percept , 2004, Trends in Cognitive Sciences.

[20]  S. Nishida,et al.  A common perceptual temporal limit of binding synchronous inputs across different sensory attributes and modalities , 2010, Proceedings of the Royal Society B: Biological Sciences.

[21]  D. Knill Robust cue integration: a Bayesian model and evidence from cue-conflict studies with stereoscopic and figure cues to slant. , 2007, Journal of vision.

[22]  R. Blake,et al.  Visual features that vary together over time group together over space , 1998, Nature Neuroscience.

[23]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[24]  Marc O. Ernst,et al.  A Bayesian view on multimodal cue integration , 2006 .

[25]  M. Wertheimer Untersuchungen zur Lehre von der Gestalt. II , 1923 .

[26]  Ulrik R. Beierholm,et al.  Causal inference in perception , 2010, Trends in Cognitive Sciences.