A Framework for Depth Estimation and Relative Localization of Ground Robots using Computer Vision

The 3D depth estimation and relative pose estimation problem within a decentralized architecture is a challenging problem that arises in missions that require coordination among multiple vision-controlled robots. The depth estimation problem aims at recovering the 3D information of the environment. The relative localization problem consists of estimating the relative pose between two robots, by sensing each other’s pose or sharing information about the perceived environment. Most solutions for these problems use a set of discrete data without taking into account the chronological order of the events. This paper builds on recent results on continuous estimation to propose a framework that estimates the depth and relative pose between two non-holonomic vehicles. The basic idea consists in estimating the depth of the points by explicitly considering the dynamics of the camera mounted on a ground robot, and feeding the estimates of 3D points observed by both cameras in a filter that computes the relative pose between the robots. We evaluate the convergence for a set of simulated scenarios and show experimental results validating the proposed framework.

[1]  Roland Siegwart,et al.  Voxblox: Incremental 3D Euclidean Signed Distance Fields for on-board MAV planning , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[2]  António Manuel Santos Pascoal,et al.  A Decentralized Strategy for Multirobot Sampling/Patrolling: Theory and Experiments , 2015, IEEE Transactions on Control Systems Technology.

[3]  David Nistér,et al.  An efficient solution to the five-point relative pose problem , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[4]  Binoy Pinto,et al.  Speeded Up Robust Features , 2011 .

[5]  Hao Su,et al.  A Point Set Generation Network for 3D Object Reconstruction from a Single Image , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Bruce Randall Donald,et al.  Moving furniture with teams of autonomous robots , 1995, Proceedings 1995 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human Robot Interaction and Cooperative Robots.

[7]  Riccardo Spica,et al.  Active decentralized scale estimation for bearing-based localization , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Gary R. Bradski,et al.  ORB: An efficient alternative to SIFT or SURF , 2011, 2011 International Conference on Computer Vision.

[9]  Daniel Cremers,et al.  Direct Sparse Odometry , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Friedrich Fraundorfer,et al.  Visual Odometry Part I: The First 30 Years and Fundamentals , 2022 .

[11]  Riccardo Spica,et al.  A framework for active estimation: Application to Structure from Motion , 2013, 52nd IEEE Conference on Decision and Control.

[12]  Hongdong Li,et al.  Efficient Computation of Relative Pose for Multi-camera Systems , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  J. M. M. Montiel,et al.  ORB-SLAM: A Versatile and Accurate Monocular SLAM System , 2015, IEEE Transactions on Robotics.

[14]  Luc Van Gool,et al.  Speeded-Up Robust Features (SURF) , 2008, Comput. Vis. Image Underst..

[15]  Sonia Martínez,et al.  Coverage control for mobile sensing networks , 2002, IEEE Transactions on Robotics and Automation.

[16]  David G. Lowe,et al.  Object recognition from local scale-invariant features , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.

[17]  Kun Huang,et al.  Learning to Parse Wireframes in Images of Man-Made Environments , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[18]  Wolfram Burgard,et al.  A Probabilistic Approach to Collaborative Multi-Robot Localization , 2000, Auton. Robots.

[19]  A. Pedro Aguiar,et al.  Low-Level Active Visual Navigation: Increasing Robustness of Vision-Based Localization Using Potential Fields , 2018, IEEE Robotics and Automation Letters.

[20]  Vladlen Koltun,et al.  Deep Fundamental Matrix Estimation , 2018, ECCV.

[21]  François Chaumette,et al.  Plane estimation by active vision from point features and image moments , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[22]  G. Whelan,et al.  Cooperative search and rescue with a team of mobile robots , 1997, 1997 8th International Conference on Advanced Robotics. Proceedings. ICAR'97.

[23]  Silvio Savarese,et al.  3D-R2N2: A Unified Approach for Single and Multi-view 3D Object Reconstruction , 2016, ECCV.

[24]  J. M. M. Montiel,et al.  Indoor robot motion based on monocular images , 2001, Robotica.

[25]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[26]  François Chaumette,et al.  Coupling active depth estimation and visual servoing via a large projection operator , 2017, Int. J. Robotics Res..

[27]  Pedro Miraldo,et al.  Minimal Solvers for Mini-Loop Closures in 3D Multi-Scan Alignment , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  François Chaumette,et al.  Visual servo control. I. Basic approaches , 2006, IEEE Robotics & Automation Magazine.

[29]  Omar Tahri,et al.  Active Structure-from-Motion for 3D Straight LinesBehaviors* This work was partially supported by the Portuguese FCT grants PD/Bd/135015/2017 (through the NETSys Doctoral Program) & SFRH/BPD/111495/2015, and ISRILARSyS Strategic Funding by the FCT project PEst-OE/EEI/LA0009/2013. , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[30]  François Chaumette,et al.  Active Structure From Motion: Application to Point, Sphere, and Cylinder , 2014, IEEE Transactions on Robotics.