An Attentional Model for Autonomous Mobile Robots

The increase of applications that use autonomous robots has endowed them with a high number of sensors and actuators that are sometimes redundant. This new highly complex systems and the type of environment where they are expected to operate require them to deal with data overload and data fusion. In humans that face the same problem when sounds, images, and smells are presented to their sensors in a daily scene, a natural filter is applied: attention. Although there are many computational models that apply attentive systems to robotics, they usually are restricted to two classes of systems: 1) those that have complex biologically based attentional visual systems and 2) those that have simpler attentional mechanisms with a larger variety of sensors. This work proposes an attentional model inspired from biological systems and that supports a variety of robotics sensors. Furthermore, it discusses the possibility of using multiple sensors to define multiple features, with feature extraction modules that can handle exogenous and endogenous attentional processes. The experiments were performed in a simulated high-fidelity environment, and results have shown that the model proposed can account for detecting salient (bottom-up) and desired (top-down) stimuli according to the modeled features.

[1]  L. Itti,et al.  Modeling the influence of task on attention , 2005, Vision Research.

[2]  Minoru Asada,et al.  Visual attention control for a legged mobile robot based on information criterion , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[3]  C. Balkenius Attention, habituation and conditioning: toward a computational model , 2000 .

[4]  Angelo Cangelosi,et al.  A neural model of selective attention and object segmentation in the visual scene: An approach based on partial synchronization and star-like architecture of connections , 2009, Neural Networks.

[5]  John G. Taylor,et al.  Attention-Based Robot Control , 2003, KES.

[6]  Pierre Baldi,et al.  A bottom-up model of spatial attention predicts human error patterns in rapid scene recognition. , 2007, Journal of vision.

[7]  Michael I. Posner,et al.  Analyzing and shaping human attentional networks , 2006, Neural Networks.

[8]  Laurent Itti,et al.  Beyond bottom-up: Incorporating task-dependent influences into a computational model of spatial attention , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[9]  M. Posner,et al.  Components of visual orienting , 1984 .

[10]  Reid G. Simmons,et al.  Multimodal person tracking and attention classification , 2006, HRI '06.

[11]  Majid Nili Ahmadabadi,et al.  Cost-sensitive learning of top-down modulation for attentional control , 2009, Machine Vision and Applications.

[12]  Leslie G. Ungerleider,et al.  Mechanisms of visual attention in the human cortex. , 2000, Annual review of neuroscience.

[13]  I. THE ATTENTION SYSTEM OF THE HUMAN BRAIN , 2002 .

[14]  D. Heeger,et al.  The Normalization Model of Attention , 2009, Neuron.

[15]  Juyang Weng,et al.  Online-learning and Attention-based Approach to Obstacle Avoidance Using a Range Finder , 2007, J. Intell. Robotic Syst..

[16]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[17]  Laurent Itti,et al.  A Goal Oriented Attention Guidance Model , 2002, Biologically Motivated Computer Vision.

[18]  Simone Frintrop 5 The Visual Attention System VOCUS: Top-Down Extension , 2006 .

[19]  M. Posner,et al.  Orienting of Attention* , 1980, The Quarterly journal of experimental psychology.

[20]  T. P. Schmidt Robotics : Environmental Awareness Through Cognitive Sensor Fusion , 2010 .

[21]  John G. Taylor,et al.  Attention as a neural control system , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).

[22]  Christian Balkenius,et al.  Attention as selection-for-action: a scheme for active perception , 1999, 1999 Third European Workshop on Advanced Mobile Robots (Eurobot'99). Proceedings (Cat. No.99EX355).

[23]  R. Klein,et al.  Inhibition of return , 2000, Trends in Cognitive Sciences.

[24]  Henrik I. Christensen,et al.  Pay Attention When Selecting Features , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[25]  Jannik Fritsch,et al.  A multi-modal object attention system for a mobile robot , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[26]  M. Carandini,et al.  Normalization as a canonical neural computation , 2013, Nature Reviews Neuroscience.

[27]  Marisa Carrasco,et al.  When sustained attention impairs perception , 2006, Nature Neuroscience.

[28]  Helgi Páll Helgason,et al.  General Attention Mechanism for Artificial Intelligence Systems , 2013 .

[29]  R. F. D. Lima Compreendendo os Mecanismos Atencionais , 2005 .

[30]  John G. Taylor,et al.  Combining Attention and Value Maps , 2005, ICANN.

[31]  Laurent Itti,et al.  Applying computational tools to predict gaze direction in interactive visual environments , 2008, TAP.

[32]  H. Pashler Divided attention: Storing and classifying briefly presented objects , 1994, Psychonomic bulletin & review.

[33]  C. Gilbert,et al.  Brain States: Top-Down Influences in Sensory Processing , 2007, Neuron.

[34]  George K. I. Mann,et al.  An Object-Based Visual Attention Model for Robotic Applications , 2010, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[35]  John G. Taylor,et al.  Attention-driven artificial agents , 2003 .

[36]  S. Wermter,et al.  Knowing what and where: a computational model for visual attention , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[37]  Simone Frintrop,et al.  An Attentive, Multi-modal Laser "Eye" , 2003, ICVS.