Perceptual Information Fusion in Humans and Machines

Humans perceive the world through different perceptual modalities, which are processed in the brain by modality-specific areas and structures. However, there also exist multimodal neurons and areas, specialized in integrating perceptual information to enhance or suppress brain response. The particular way the human brain fuses crossmodal (or multimodal) perceptual information manifests itself first in behavioural studies. These crossmodal interactions are widely explored in some modalities, especially for auditory and visual input, and less explored for other modalities, like taste and olfaction, yet it is known that these effects can occur with any two modalities. The integration of sensory data is an important research area in computer science, and stands to benefit from the studies into the brain function; many biological processes serve as models for computer algorithms. On the other hand, computer models of sensor integration are built on mathematical principles, and provide normative insights into the functioning of the brain. This paper surveys the psychological and neurological findings pertaining to human multisensor fusion, followed by a brief review of the relevant computer science terminology and modeling approaches. The potential of an interdisciplinary approach to information fusion encompassing neuroscience, psychology and computer science have recently been recognized, and a multidisciplinary workshop on biologically inspired information fusion was organized to bring researchers together to determine a common agenda. The conclusion summarizes the agenda of research outlined at the workshop and attempts to raise research questions for the future.

[1]  Albert Ali Salah,et al.  A Selective Attention-Based Method for Visual Pattern Recognition with Application to Handwritten Digit Recognition and Face Recognition , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  R. Sekuler,et al.  Sound alters visual motion perception , 1997, Nature.

[3]  Antoinette T. Gesi,et al.  Bimodal speech perception: an examination across languages , 1993 .

[4]  E. Rolls,et al.  Taste‐olfactory convergence, and the representation of the pleasantness of flavour, in the human brain , 2003, The European journal of neuroscience.

[5]  Belur V. Dasarathy,et al.  Sensor fusion potential exploitation-innovative architectures and illustrative applications , 1997, Proc. IEEE.

[6]  M A Schmuckler,et al.  The McGurk effect in infants. , 1995, Perception & psychophysics.

[7]  Marty G. Woldorff,et al.  Selective Attention and Multisensory Integration: Multiple Phases of Effects on the Evoked Brain Activity , 2005, Journal of Cognitive Neuroscience.

[8]  T. Valentine The Quarterly Journal of Experimental Psychology Section A: Human Experimental Psychology a Unified Account of the Effects of Distinctiveness, Inversion, and Race in Face Recognition , 2022 .

[9]  Hugh F. Durrant-Whyte,et al.  Sensor Models and Multisensor Integration , 1988, Int. J. Robotics Res..

[10]  Paul Bertelson,et al.  The effect of a textured visual field on modality dominance in a ventriloquism situation , 1976 .

[11]  Pascal Vasseur,et al.  Introduction to Multisensor Data Fusion , 2005, The Industrial Information Technology Handbook.

[12]  B E Stein,et al.  Effects of cooling somatosensory cortex on response properties of tactile cells in the superior colliculus. , 1986, Journal of neurophysiology.

[13]  H. McGurk,et al.  Hearing lips and seeing voices , 1976, Nature.

[14]  S. Shimojo,et al.  Illusions: What you see is what you hear , 2000, Nature.

[15]  I. Rock,et al.  Vision and touch. , 1967, Scientific American.

[16]  O. Bock,et al.  Contribution of retinal versus extraretinal signals towards visual localization in goal-directed movements , 2004, Experimental Brain Research.

[17]  Mikko Sams,et al.  Factors influencing audiovisual fission and fusion illusions. , 2004, Brain research. Cognitive brain research.

[18]  P. Bertelson,et al.  The psychology of multimodal perception , 2004 .

[19]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[20]  C. Fowler,et al.  Listening with eye and hand: cross-modal contributions to speech perception. , 1991, Journal of experimental psychology. Human perception and performance.

[21]  Sidney S. Simon,et al.  Merging of the Senses , 2008, Front. Neurosci..

[22]  M. Giard,et al.  Auditory-Visual Integration during Multimodal Object Recognition in Humans: A Behavioral and Electrophysiological Study , 1999, Journal of Cognitive Neuroscience.

[23]  Berk Gökberk,et al.  Rank-based decision fusion for 3D shape-based face recognition , 2005, SIU 2005.

[24]  P. Dalton,et al.  Flavor processing: perceptual and cognitive factors in multi-modal integration. , 2005, Chemical senses.

[25]  J. Driver Enhancement of selective listening by illusory mislocation of speech sounds due to lip-reading , 1996, Nature.

[26]  M. Wallace,et al.  Converging influences from visual, auditory, and somatosensory cortices onto output neurons of the superior colliculus. , 1993, Journal of neurophysiology.

[27]  H. Pick,et al.  Visual capture produced by prism spectacles , 1965 .

[28]  A. Gilbert,et al.  Cross-modal correspondence between vision and olfaction: the color of smells. , 1996, The American journal of psychology.

[29]  Edmund T Rolls,et al.  Smell, taste, texture, and temperature multimodal representations in the brain, and their relevance to the control of appetite. , 2004, Nutrition reviews.

[30]  S Shimojo,et al.  Attentional Modulation in Perception of Visual Motion Events , 1998, Perception.

[31]  Robin R. Murphy,et al.  Biological and cognitive foundations of intelligent sensor fusion , 1996, IEEE Trans. Syst. Man Cybern. Part A.

[32]  D. Vaitl,et al.  Utilizing the ventriloquism-effect to investigate audio-visual binding , 2007, Neuropsychologia.

[33]  E. Bullmore,et al.  Activation of auditory cortex during silent lipreading. , 1997, Science.

[34]  D. H. Warren,et al.  Immediate perceptual response to intersensory discrepancy. , 1980, Psychological bulletin.

[35]  P. Bertelson Chapter 14 Ventriloquism: A case of crossmodal perceptual grouping , 1999 .

[36]  C. Spence,et al.  Attention and the crossmodal construction of space , 1998, Trends in Cognitive Sciences.

[37]  G. Calvert Crossmodal processing in the human brain: insights from functional neuroimaging studies. , 2001, Cerebral cortex.

[38]  Mark T. Wallace,et al.  Crossmodal spatial interactions in subcortical and cortical circuits , 2004 .

[39]  D. Dubourdieu,et al.  The Color of Odors , 2001, Brain and Language.

[40]  Tony Ro,et al.  Visually induced feelings of touch , 2006, Brain Research.

[41]  V. Jousmäki,et al.  Parchment-skin illusion: sound-biased touch , 1998, Current Biology.

[42]  E I Knudsen,et al.  Visual instruction of the neural map of auditory space in the developing optic tectum. , 1991, Science.

[43]  Jonathan D. Cohen,et al.  Rubber hands ‘feel’ touch that eyes see , 1998, Nature.

[44]  E I Knudsen,et al.  Vision guides the adjustment of auditory localization in young barn owls. , 1985, Science.

[45]  D. Laird,et al.  How the consumer estimates quality by subconscious sensory impressions. , 1932 .

[46]  S. Shimojo,et al.  Visual illusion induced by sound. , 2002, Brain research. Cognitive brain research.

[47]  Berk Gökberk,et al.  Rank-based decision fusion for 3D shape-based face recognition , 2005, Proceedings of the IEEE 13th Signal Processing and Communications Applications Conference, 2005..

[48]  J. Daugman Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters. , 1985, Journal of the Optical Society of America. A, Optics and image science.

[49]  R. Dolan,et al.  The Nose Smells What the Eye Sees Crossmodal Visual Facilitation of Human Olfactory Perception , 2003, Neuron.

[50]  Alexandre Pouget,et al.  A computational neural theory of multisensory spatial representations , 2004 .