Minimalistic Vision-based Cognitive SLAM

The interest in cognitive robotics is still increasing, a major goal being to create a system which can adapt to dynamic environments and which can learn from its own experiences. We present a new cognitive SLAM architecture, but one which is minimalistic in terms of sensors and memory. It employs only one camera with pan and tilt control and three memories, without additional sensors nor any odometry. Short-term memory is an egocentric map which holds information at close range at the actual robot position. Long-term memory is used for mapping the environment and registration of encountered objects. Object memory holds features of learned objects which are used as navigation landmarks and task targets. Saliency maps are used to sequentially focus important areas for object and obstacle detection, but also for selecting directions of movements. Reinforcement learning is used to consolidate or enfeeble environmental information in long-term memory. The system is able to achieve complex tasks by executing sequences of visuomotor actions, decisions being taken by goal-detection and goal-completion tasks. Experimental results show that the system is capable of executing tasks like localizing specific objects while building a map, after which it manages to return to the start position even when new obstacles have appeared.

[1]  Ronald A. Rensink The Dynamic Representation of Scenes , 2000 .

[2]  Sebastian Thrun,et al.  FastSLAM: a factored solution to the simultaneous localization and mapping problem , 2002, AAAI/IAAI.

[3]  Srikanta Patnaik Robot cognition and navigation - an experiment with mobile robots , 2007, Cognitive Technologies.

[4]  Gordon Wyeth,et al.  Persistent Navigation and Mapping using a Biologically Inspired SLAM System , 2010, Int. J. Robotics Res..

[5]  Michael Zillich,et al.  Biologically inspired navigation on a mobile robot , 2010, 2010 IEEE International Conference on Robotics and Biomimetics.

[6]  Ludmilla Kleinmann,et al.  Learning to adapt: Cognitive architecture based on biologically inspired memory , 2011, 2011 6th IEEE Conference on Industrial Electronics and Applications.

[7]  Garrison W. Cottrell,et al.  Visual saliency model for robot cameras , 2008, 2008 IEEE International Conference on Robotics and Automation.

[8]  João M. F. Rodrigues,et al.  A vision system for detecting paths and moving obstacles for the blind , 2010 .

[9]  Palis Ratanaswasd,et al.  Cognitive control for robot task execution , 2005, ROMAN 2005. IEEE International Workshop on Robot and Human Interactive Communication, 2005..

[10]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[11]  James J. Little,et al.  Curious George: An attentive semantic robot , 2008, Robotics Auton. Syst..

[12]  João M. F. Rodrigues,et al.  Automatic hand or head gesture interface for individuals with motor impairments, senior citizens and young children , 2009 .

[13]  Pamela A. Meinert The Effect of Previous Life Experience on Cognitive Structure Changes and Knowledge Acquisition of Nursing Theory and Clinical Skills in Nontraditional Nursing Students , 2008 .

[14]  Rachid Alami,et al.  Toward Human-Aware Robot Task Planning , 2006, AAAI Spring Symposium: To Boldly Go Where No Human-Robot Team Has Gone Before.

[15]  D. M. Wilkes,et al.  TOWARD EGOCENTRIC NAVIGATION , 2002 .

[16]  João Rodrigues,et al.  Multi-scale keypoints in V1 and beyond: object segregation, scale selection, saliency maps and face detection. , 2006, Bio Systems.

[17]  Aude Oliva,et al.  Visual long-term memory has a massive storage capacity for object details , 2008, Proceedings of the National Academy of Sciences.

[18]  João M. F. Rodrigues,et al.  Disparity energy model with keypoint disparity validation , 2011 .

[19]  João M. F. Rodrigues,et al.  Multi-scale lines and edges in V1 and beyond: Brightness, object categorization and recognition, and consciousness , 2009, Biosyst..