Embedded and real-time architecture for bio-inspired vision-based robot navigation

AbstractA recent trend in several robotics tasks is to consider vision as the primary sense to perceive the environment or to interact with humans. Therefore, vision processing becomes a central and challenging matter for the design of real-time control architectures. We follow in this paper a biological inspiration to propose a real-time and embedded control system relying on visual attention to learn specific actions in each place recognized by our robot. Faced with a performance challenge, the attentional model allows to reduce vision processing to a few regions of the visual field. However, the computational complexity of the visual chain remains an issue for a processing system embedded onto an indoor robot. That is why we propose as the first part of our system, a full-hardware architecture prototyped onto reconfigurable devices to detect salient features at the camera frequency. The second part learns continuously these features in order to implement specific robotics tasks. This neural control layer is implemented as embedded software making the robot fully autonomous from a computation point of view. The integration of such a system onto the robot enables not only to accelerate the frame rate of the visual processing, to relieve the control architecture but also to compress the data-flow at the output of the camera, thus reducing communication and energy consumption. We present in this paper the complete embedded sensorimotor architecture and the experimental setup. The presented results demonstrate its real-time behavior in vision-based navigation tasks.

[1]  Heinz Hügli,et al.  Robot self-localization using visual attention , 2005, 2005 International Symposium on Computational Intelligence in Robotics and Automation.

[2]  Philippe Gaussier,et al.  Neurobiologically Inspired Mobile Robot Navigation and Planning , 2007, Frontiers in neurorobotics.

[3]  Cordelia Schmid,et al.  Scale & Affine Invariant Interest Point Detectors , 2004, International Journal of Computer Vision.

[4]  Philippe Gaussier,et al.  Distributed real time neural networks in interactive complex systems , 2008, CSTST.

[5]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[6]  R. Passingham The hippocampus as a cognitive map J. O'Keefe & L. Nadel, Oxford University Press, Oxford (1978). 570 pp., £25.00 , 1979, Neuroscience.

[7]  Yann LeCun,et al.  An FPGA-based stream processor for embedded real-time vision with Convolutional Networks , 2009, 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops.

[8]  Philippe Gaussier,et al.  The visual homing problem: An example of robotics/biology cross fertilization , 2000, Robotics Auton. Syst..

[9]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[10]  M. Goodale,et al.  Separate visual pathways for perception and action , 1992, Trends in Neurosciences.

[11]  François Berry,et al.  FPGA-based real time extraction of visual features , 2012, 2012 IEEE International Symposium on Circuits and Systems.

[12]  T. Collett,et al.  Multiple stored views and landmark guidance in ants , 1998, Nature.

[13]  Michael Recce,et al.  A model of hippocampal function , 1994, Neural Networks.

[14]  Zhiguo Cao,et al.  A real-time embedded architecture for SIFT , 2013, J. Syst. Archit..

[15]  Philippe Gaussier,et al.  Visual navigation in an open environment without map , 1997, Proceedings of the 1997 IEEE/RSJ International Conference on Intelligent Robot and Systems. Innovative Robotics for Real-World Applications. IROS '97.

[16]  Simone Frintrop,et al.  Attentional Landmarks and Active Gaze Control for Visual SLAM , 2008, IEEE Transactions on Robotics.

[17]  Frank W. Grasso,et al.  Biomimetic robot lobster performs chemo-orientation in turbulence using a pair of spatially separated sensors: Progress and challenges , 2000, Robotics Auton. Syst..

[18]  Laurent Itti,et al.  Biologically Inspired Mobile Robot Vision Localization , 2009, IEEE Transactions on Robotics.

[19]  Philippe Gaussier,et al.  PerAc: A neural architecture to control artificial animals , 1995, Robotics Auton. Syst..

[20]  Philippe Gaussier,et al.  Biologically inspired neural networks for spatio-temporal planning in robotic navigation tasks , 2011, 2011 IEEE International Conference on Robotics and Biomimetics.

[21]  Eduardo Marques,et al.  An Embedded Multi-camera System for Simultaneous Localization and Mapping , 2006, ARC.

[22]  B. Kolb,et al.  The Cerebral cortex of the rat , 1990 .

[23]  P. Gaussiera,et al.  The visual homing problem : An example of robotics / biology cross fertilization , 1999 .

[24]  Philippe Gaussier,et al.  Robustness of Visual Place Cells in Dynamic Indoor and Outdoor Environment , 2006 .

[25]  Philippe Gaussier,et al.  Perception as a Dynamical Sensori-Motor Attraction Basin , 2005, ECAL.

[26]  L. Nadel,et al.  The Hippocampus as a Cognitive Map , 1978 .

[27]  永福 智志 The Organization of Learning , 2005, Journal of Cognitive Neuroscience.

[28]  Niccolò Battezzati,et al.  SURF algorithm in FPGA: A novel architecture for high demanding industrial applications , 2012, 2012 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[29]  Dana H. Ballard,et al.  Animate Vision , 1991, Artif. Intell..

[30]  E. Culurciello,et al.  NeuFlow: Dataflow vision processing system-on-a-chip , 2012, 2012 IEEE 55th International Midwest Symposium on Circuits and Systems (MWSCAS).

[31]  Benoît Miramond,et al.  Using High-Level RTOS Models for HW/SW Embedded Architecture Exploration: Case Study on Mobile Robotic Vision , 2008, EURASIP J. Embed. Syst..

[32]  H. Evans The Study of Instinct , 1952 .

[33]  John K. Tsotsos Analyzing vision at the complexity level , 1990, Behavioral and Brain Sciences.

[34]  T. S. Collett,et al.  Landmark learning in bees , 1983, Journal of comparative physiology.

[35]  Henrik I. Christensen,et al.  Computational visual attention systems and their cognitive foundations: A survey , 2010, TAP.

[36]  James L. Crowley,et al.  Fast Computation of Scale Normalised Gaussian Receptive Fields , 2003, Scale-Space.

[37]  D. Zipser,et al.  Biologically plausible models of place recognition and goal location , 1986 .

[38]  Ioannis Papaefstathiou,et al.  Fast and Efficient FPGA-Based Feature Detection Employing the SURF Algorithm , 2010, 2010 18th IEEE Annual International Symposium on Field-Programmable Custom Computing Machines.

[39]  William Rowan,et al.  The Study of Instinct , 1953 .

[40]  J. O’Keefe,et al.  Neuronal computations underlying the firing of place cells and their role in navigation , 1996, Hippocampus.

[41]  Tony Lindeberg,et al.  Feature Detection with Automatic Scale Selection , 1998, International Journal of Computer Vision.

[42]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[43]  G. Humphreys,et al.  Computational models of visual selective attention: A review , 2005 .

[44]  Antonio Torralba,et al.  Top-down control of visual attention in object detection , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).

[45]  Philippe Gaussier,et al.  From view cells and place cells to cognitive map learning: processing stages of the hippocampal system , 2002, Biological Cybernetics.

[46]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[47]  Gundolf Kiefer,et al.  Flex-SURF: A Flexible Architecture for FPGA-Based Robust Feature Extraction for Optical Tracking Systems , 2010, 2010 International Conference on Reconfigurable Computing and FPGAs.

[48]  A. Treisman,et al.  A feature-integration theory of attention , 1980, Cognitive Psychology.

[49]  George A. Constantinides,et al.  A Parallel Hardware Architecture for Scale and Rotation Invariant Feature Detection , 2008, IEEE Transactions on Circuits and Systems for Video Technology.