The "puzzle" of sensory perception: putting together multisensory information

For perceiving the environment our brain uses multiple sourcesof sensory information derived from several different modalities,including vision, touch and audition. The question how informationderived from these different sensory modalities converges in thebrain to form a coherent and robust percept is central tounderstanding the process of perception. My main research interestis the study of human perception focusing on multimodal integrationand visual-haptic interaction. For this, I use quantitativecomputational/statistical models together with psychophysical andneuropsychological methods. A desirable goal for the perceptual system is to maximize thereliability of the various perceptual estimates. From a statisticalviewpoint the optimal strategy for achieving this goal is tointegrate all available sensory information. This may be done usinga "maximum-likelihood-estimation" (MLE) strategy. Then the combinedpercept will be a weighted average across the individual estimateswith weights that are proportional to their reliabilities. In a recent study we could show that humans actually integratevisual and haptic information in such a statistically optimalfashion (Ernst & Banks, Nature, 2002). Others have nowdemonstrated that this finding is true not only for the integrationacross vision and touch, but also for the integration ofinformation across and within other modalities, such as audition orvision. This suggests that maximum-likelihood-estimation is aneffective and widely used strategy exploited by the perceptualsystem. By integrating sensory information the brain may or may notloose access to the individual input signals feeding into theintegrated percept. The degree to which the original information isstill accessible defines the strength of coupling between thesignals. We found that the strengths of coupling is varyingdepending on the set of signals used; e.g. strong coupling forstereo and texture signals to slant and weak coupling for visualand haptic signals to size (Hillis, Ernst, Banks, & Landy,Science, 2002). As suggested by one of our recent learning studies,the strength of coupling, which can be modeled using Bayesianstatistics, seems to depend on the natural statisticalco-occurrence between signals (Jäkel & Ernst, inprep.) Important precondition for integrating signals is to know whichsignals derived from the different modalities belong together andhow reliable these are. Recently we could show that touch can teachthe visual modality how to interpret its signals and theirreliabilities. More specifically, we could show that by exploitingtouch we can alter visual perception of slant (Ernst, Banks &Bulthoff, Nature Neuroscience, 2000). This finding contributes to avery old debate postulating that we only perceive the world becauseof our interactions with the environment. Similarly, in one of ourlatest studies we could show that experience can change theso-called "light-from-above" prior. Prior knowledge is essentialfor the interpretation of sensory signals during perception.Consequently, with the prior change we introduced a change in theperception of shape (Adams, Graf & Ernst, Nature Neuroscience,2004). Integration is only sensible if the information sources carryredundant information. If the information sources arecomplementary, different combination strategies have to beexploited. Complementation of cross-modal information wasdemonstrated in a recent study investigating visual-haptic shapeperception (Newell, Ernst, Tjan, & Bulthoff, PsychologicalScience, 2001).

[1]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[2]  Knut Drewing,et al.  Cue Integration in the Haptic Perception of Virtual Shapes , 2003 .

[3]  M. Ernst Learning to integrate arbitrary signals from vision and touch. , 2007, Journal of vision.

[4]  M. Ernst,et al.  Optimal integration of shape information from vision and touch , 2007, Experimental Brain Research.

[5]  Mo Ernst A Bayesian view on multimodal integration , 2005 .

[6]  Martin Buss,et al.  Construction and first evaluation of a newly developed tactile Shear Force Display , 2004 .

[7]  G. Reibnegger,et al.  Calorimetry of archaeal tetraether lipid—indication of a novel metastable thermotropic phase in the main phospholipid from Thermoplasma acidophilum cultured at 59°C , 1998 .

[8]  Heinrich H. Bülthoff,et al.  Touch can change visual slant perception , 2000, Nature Neuroscience.

[9]  S. Gepshtein,et al.  The combination of vision and touch depends on spatial proximity. , 2005, Journal of vision.

[10]  M. Ernst,et al.  Experience can change the 'light-from-above' prior , 2004, Nature Neuroscience.

[11]  Heinrich H. Bülthoff,et al.  Cross-modal perception of actively explored objects , 2003 .

[12]  H. Bülthoff,et al.  Merging the senses into a robust percept , 2004, Trends in Cognitive Sciences.

[13]  M. Ernst,et al.  Feeling what you hear: auditory signals can modulate tactile tap perception , 2005, Experimental Brain Research.

[14]  Thomas V. Wiecki,et al.  Material Properties Determine How we Integrate Shape Signals in Active Touch , 2005 .

[15]  M. Ernst,et al.  First evaluation of a novel tactile Shear Force Display , 2004 .

[16]  Jean-Pierre Bresciani,et al.  Vision and touch are automatically integrated for the perception of sequences of events. , 2006, Journal of vision.

[17]  Frank Jäkel,et al.  Learning to Combine Arbitrary Signals from Vision and Touch , 2003 .

[18]  M. Ernst,et al.  Signal reliability modulates auditory–tactile integration for event counting , 2007, Neuroreport.

[19]  Martin Buss,et al.  Integration of Kinesthetic and Tactile Display: A Modular Design Concept , 2006 .

[20]  James M. Hillis,et al.  Combining Sensory Information: Mandatory Fusion Within, but Not Between, Senses , 2002, Science.

[21]  Marc O. Ernst Psychophysikalische Untersuchungen zur visuomotorischen Integration beim Menschen: visuelle und haptische Wahrnehmung virtueller und realer Objekte. , 2001 .

[22]  H. Bülthoff,et al.  Viewpoint Dependence in Visual and Haptic Object Recognition , 2001, Psychological science.

[23]  Peter Carr,et al.  Invited talks , 2005, NUSOD '05. Proceedings of the 5th International Conference on Numerical Simulation of Optoelectronic Devices, 2005..

[24]  Laurence Maloney,et al.  Combining sensory information to improve visualization , 2002, VIS '02.

[25]  Marc O. Ernst,et al.  Effect of attention on multimodal cue integration , 2004 .

[26]  M. Ernst,et al.  Multisensory recognition of actively explored objects. , 2007, Canadian journal of experimental psychology = Revue canadienne de psychologie experimentale.

[27]  Marc O. Ernst,et al.  Bimanual Size Estimation: No Automatic Integration of Information across the Hands , 2004 .

[28]  M. Ernst,et al.  Focus cues affect perceived depth. , 2005, Journal of vision.

[29]  Marc O. Ernst Optimal Integration of Multimodal Information: Conditions and Limits , 2004 .

[30]  Johan Wagemans,et al.  Texture and haptic cues in slant discrimination: reliability-based cue weighting without statistically optimal cue combination. , 2005, Journal of the Optical Society of America. A, Optics, image science, and vision.