A nonparametric learning approach to range sensing from omnidirectional vision

We present a novel approach to estimating depth from single omnidirectional camera images by learning the relationship between visual features and range measurements available during a training phase. Our model not only yields the most likely distance to obstacles in all directions, but also the predictive uncertainties for these estimates. This information can be utilized by a mobile robot to build an occupancy grid map of the environment or to avoid obstacles during exploration-tasks that typically require dedicated proximity sensors such as laser range finders or sonars. We show in this paper how an omnidirectional camera can be used as an alternative to such range sensors. As the learning engine, we apply Gaussian processes, a nonparametric approach to function regression, as well as a recently developed extension for dealing with input-dependent noise. In practical experiments carried out in different indoor environments with a mobile robot equipped with an omnidirectional camera system, we demonstrate that our system is able to estimate range with an accuracy comparable to that of dedicated sensors based on sonar or infrared light.

[1]  Tomás Pajdla,et al.  Structure from motion with wide circular field of view cameras , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  Bernd Freisleben,et al.  Using depth features to retrieve monocular video shots , 2007, CIVR '07.

[3]  Stephen Grossberg,et al.  Laminar cortical mechanisms for the perception of slanted and curved 3-D surfaces and their 2-D pictorical projections , 2010 .

[4]  Sebastian Thrun,et al.  Self-supervised Monocular Road Detection in Desert Terrain , 2006, Robotics: Science and Systems.

[5]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[6]  Wolfram Burgard,et al.  Visual Bearing-Only Simultaneous Localization and Mapping with Improved Feature Matching , 2007, AMS.

[7]  E. R. Davies,et al.  Machine vision - theory, algorithms, practicalities , 2004 .

[8]  Ashutosh Saxena,et al.  3-D Depth Reconstruction from a Single Still Image , 2007, International Journal of Computer Vision.

[9]  Takeshi Ohashi,et al.  Obstacle avoidance and path planning for humanoid robots using stereo vision , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[10]  Dit-Yan Yeung,et al.  Robust locally linear embedding , 2006, Pattern Recognit..

[11]  Wolfram Burgard,et al.  Monocular range sensing: A non-parametric learning approach , 2008, 2008 IEEE International Conference on Robotics and Automation.

[12]  Volker Tresp,et al.  Mixtures of Gaussian Processes , 2000, NIPS.

[13]  James J. Little,et al.  /spl sigma/SLAM: stereo vision SLAM using the Rao-Blackwellised particle filter and a novel mixture proposal distribution , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[14]  Emanuele Menegatti,et al.  Omnidirectional vision scan matching for robot localization in dynamic environments , 2006, IEEE Transactions on Robotics.

[15]  Christian Laugier,et al.  Autonomous Navigation in Dynamic Environments , 2007 .

[16]  Luc Van Gool,et al.  Speeded-Up Robust Features (SURF) , 2008, Comput. Vis. Image Underst..

[17]  Wolfram Burgard,et al.  Gas Distribution Modeling using Sparse Gaussian Process Mixture Models , 2008, Robotics: Science and Systems.

[18]  Wolfram Burgard,et al.  Map learning and high-speed navigation in RHINO , 1998 .

[19]  Alexei A. Efros,et al.  Recovering Surface Layout from an Image , 2007, International Journal of Computer Vision.

[20]  Carl E. Rasmussen,et al.  Learning Depth from Stereo , 2004, DAGM-Symposium.

[21]  Antonio Torralba,et al.  Depth Estimation from Image Structure , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[22]  Christopher Hunt,et al.  Notes on the OpenSURF Library , 2009 .

[23]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[24]  Stefano Soatto,et al.  A geometric approach to shape from defocus , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Wolfram Burgard,et al.  Gaussian Beam Processes: A Nonparametric Bayesian Measurement Model for Range Finders , 2007, Robotics: Science and Systems.

[26]  Wolfram Burgard,et al.  Mobile Robot Map Learning from Range Data in Dynamic Environments , 2007 .

[27]  S. T. Buckland,et al.  An Introduction to the Bootstrap. , 1994 .

[28]  Ashutosh Saxena,et al.  High speed obstacle avoidance using monocular vision and reinforcement learning , 2005, ICML.

[29]  Feng Han,et al.  Bayesian reconstruction of 3D shapes and scenes from a single image , 2003, First IEEE International Workshop on Higher-Level Knowledge in 3D Modeling and Motion Analysis, 2003. HLK 2003..

[30]  Yun-Su Ha,et al.  Environmental map building for a mobile robot using infrared range-finder sensors , 2004, Adv. Robotics.

[31]  John F. Canny,et al.  A Computational Approach to Edge Detection , 1986, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[33]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[34]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[35]  James J. Little,et al.  Autonomous vision-based exploration and mapping using hybrid maps and Rao-Blackwellised particle filters , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[36]  Hans P. Moravec,et al.  High resolution maps from wide angle sonar , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[37]  Honglak Lee,et al.  Automatic Single-Image 3d Reconstructions of Indoor Manhattan World Scenes , 2007, ISRR.

[38]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .