A cognitive approach to multimodal attention

An adaptive attention mechanism is a requirement when an autonomous robot has to deal with real world environments. In this paper we present a novel cognitive architecture which enables integrated and efficient filtering of multiple modality sensory information. The proposed attention mechanism is based on contexts that determine what sensorimotor data is relevant to the current situation. These contexts are used as a mean to adaptively select constrained cognitive focus within the vast multimodal sensory space. In this framework, the focus of attention can be directed to meaningful complex percepts, thus allowing the implementation of higher cognitive capabilities. Sonar, contact, and visual sensory modalities have been used in the perception process, and the motor capability of the physical agent is provided by a differential wheel drive system. The testing of this artificial attention approach, carried out initially in the domain of counterpart recognition and chasing, has demonstrated both a great decrease in computation power requirements and ease of multimodal integration for cognitive representations.

[1]  B. Baars A cognitive theory of consciousness , 1988 .

[2]  G. Rizzolatti,et al.  Space coding by premotor cortex , 2004, Experimental Brain Research.

[3]  C. Spence,et al.  Multisensory Integration: Maintaining the Perception of Synchrony , 2003, Current Biology.

[4]  Araceli Sanchis,et al.  Applying machine consciousness models in autonomous situated agents , 2008, Pattern Recognit. Lett..

[5]  Daniel Senkowski,et al.  Good times for multisensory integration: Effects of the precision of temporal synchrony as revealed by gamma-band oscillations , 2007, Neuropsychologia.

[6]  Araceli Sanchis,et al.  Modeling Consciousness for Autonomous Robot Exploration , 2007, IWINAC.

[7]  B. Wandell Foundations of vision , 1995 .

[8]  Senén Barro,et al.  Landmark detection in mobile robotics using fuzzy temporal rules , 2004, IEEE Transactions on Fuzzy Systems.

[9]  M. Giard,et al.  Auditory-Visual Integration during Multimodal Object Recognition in Humans: A Behavioral and Electrophysiological Study , 1999, Journal of Cognitive Neuroscience.

[10]  Gary R. Bradski,et al.  Peripheral-Foveal Vision for Real-time Object Recognition and Tracking in Video , 2007, IJCAI.

[11]  Ren C. Luo,et al.  Multisensor integration and fusion in intelligent systems , 1989, IEEE Trans. Syst. Man Cybern..

[12]  Daniel L. Schacter,et al.  Brain regions associated with retrieval of structurally coherent visual information , 1995, Nature.

[13]  Karl J. Friston,et al.  A direct demonstration of functional specialization in human visual cortex , 1991, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[14]  Sebastian Thrun,et al.  Probabilistic Algorithms in Robotics , 2000, AI Mag..

[15]  A. Revonsuo,et al.  Binding and Consciousness , 1999, Consciousness and Cognition.

[16]  Simone Frintrop,et al.  Visual Attention for Object Recognition in Spatial 3D Data , 2004, WAPCV.

[17]  A. Pouget,et al.  Reference frames for representing visual and tactile locations in parietal cortex , 2005, Nature Neuroscience.

[18]  Peter J. Burt,et al.  Attention mechanisms for vision in a dynamic world , 1988, [1988 Proceedings] 9th International Conference on Pattern Recognition.