Combining local and global visual information in context-based neurorobotic navigation

In robotic navigation, biologically inspired localization models have often exhibited interesting features and proven to be competitive with other solutions in terms of adaptability and performance. In general, place recognition systems rely on global or local visual descriptors; or both. In this paper, we propose a model of context-based place cells combining these two information. Global visual features are extracted to represent visual contexts. Based on the idea of global precedence, contexts drive a more refined recognition level which has local visual descriptors as an input. We evaluate this model on a robotic navigation dataset that we recorded in the outdoors. Thus, our contribution is twofold: 1) a bio-inspired model of context-based place recognition using neural networks; and 2) an evaluation assessing its suitability for applications on real robot by comparing it to 4 other architectures - 2 variants of the model and 2 stacking-based solutions - in terms of performance and computational cost. The context-based model gets the highest score based on the three metrics we consider - or is second to one of its variants. Moreover, a key feature makes the computational cost constant over time while it increases with the other methods. These promising results suggest that this model should be a good candidate for a robust place recognition in wide environments.

[1]  D. Navon Forest before trees: The precedence of global features in visual perception , 1977, Cognitive Psychology.

[2]  Philippe Gaussier,et al.  From view cells and place cells to cognitive map learning: processing stages of the hippocampal system , 2002, Biological Cybernetics.

[3]  Philippe Gaussier,et al.  Distributed real time neural networks in interactive complex systems , 2008, CSTST.

[4]  Luc Van Gool,et al.  Speeded-Up Robust Features (SURF) , 2008, Comput. Vis. Image Underst..

[5]  David M. Smith Provided for Non-commercial Research and Educational Use Only. Not for Reproduction, Distribution or Commercial Use. the Hippocampus, Context Processing and Episodic Memory , 2022 .

[6]  Kathryn J Jeffery,et al.  Heterogeneous Modulation of Place Cell Firing by Changes in Context , 2003, The Journal of Neuroscience.

[7]  Lagarde Matthieu,et al.  Distributed real time neural networks in interactive complex systems , 2008, CSTST 2008.

[8]  Michael Milford,et al.  Condition-invariant, top-down visual place recognition , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[9]  M. Bar Visual objects in context , 2004, Nature Reviews Neuroscience.

[10]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[11]  Abdul Rahman Ramli,et al.  Integration of Global and Local Salient Features for Scene Modeling in Mobile Robot Applications , 2013, Journal of Intelligent & Robotic Systems.

[12]  Philippe Gaussier,et al.  Robustness Study of a Multimodal Compass Inspired from HD-Cells and Dynamic Neural Fields , 2014, SAB.

[13]  Gordon Wyeth,et al.  Mapping a Suburb With a Single Camera Using a Biologically Inspired SLAM System , 2008, IEEE Transactions on Robotics.

[14]  Philippe Gaussier,et al.  Interest of Spatial Context for a Place Cell Based Navigation Model , 2008, SAB.

[15]  Koen E. A. van de Sande,et al.  A comparison of color features for visual concept classification , 2008, CIVR '08.

[16]  David M. Smith Chapter 4.4 The hippocampus, context processing and episodic memory , 2008 .

[17]  Philippe Gaussier,et al.  From self-assessment to frustration, a small step toward autonomy in robotic navigation , 2013, Front. Neurorobot..

[18]  Philippe Gaussier,et al.  Orientation system in Robots: Merging Allothetic and Idiothetic Estimations , 2007 .

[19]  Antonio Torralba,et al.  Building the gist of a scene: the role of global image features in recognition. , 2006, Progress in brain research.

[20]  Matthew B. Blaschko,et al.  Combining Local and Global Image Features for Object Class Recognition , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[21]  Ke Chen,et al.  Computational cognitive models of spatial memory in navigation space: A review , 2015, Neural Networks.

[22]  Rachid Deriche,et al.  Using Canny's criteria to derive a recursively implemented optimal edge detector , 1987, International Journal of Computer Vision.

[23]  Laurent Itti,et al.  Biologically Inspired Mobile Robot Vision Localization , 2009, IEEE Transactions on Robotics.

[24]  Philippe Gaussier,et al.  PerAc: A neural architecture to control artificial animals , 1995, Robotics Auton. Syst..

[25]  Philippe Gaussier,et al.  Robustness of Visual Place Cells in Dynamic Indoor and Outdoor Environment , 2006 .

[26]  M. Chun,et al.  Contextual cueing of visual attention , 2022 .