Autonomous 3-D Reconstruction, Mapping, and Exploration of Indoor Environments With a Robotic Arm

We propose a novel information gain metric that combines hand-crafted and data-driven metrics to address the next best view problem for autonomous 3-D mapping of unknown indoor environments. For the hand-crafted metric, we propose an entropy-based information gain that accounts for the previous view points to avoid the camera to revisit the same location and to promote the motion toward unexplored or occluded areas. However, for the learnt metric, we adopt a convolutional neural network (CNN) architecture and formulate the problem as a classification problem. The CNN takes the current depth image as input and outputs the motion direction that suggests the largest unexplored surface. We train and test the CNN using a new synthetic dataset based on the SUNCG dataset. The learnt motion direction is then combined with the proposed hand-crafted metric to help handle situations where using only the hand-crafted metric tends to face ambiguities. We finally evaluate the autonomous paths over several real and synthetic indoor scenes including complex industrial and domestic settings and prove that our combined metric is able to further improve the exploration coverage compared to using only the proposed hand-crafted metric.

[1]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[2]  Paolo Fiorini,et al.  Recognition self-awareness for active object recognition on depth images , 2018, BMVC.

[3]  Wolfram Burgard,et al.  OctoMap: an efficient probabilistic 3D mapping framework based on octrees , 2013, Autonomous Robots.

[4]  Vladlen Koltun,et al.  Open3D: A Modern Library for 3D Data Processing , 2018, ArXiv.

[5]  Pat Hanrahan,et al.  Submodular Trajectory Optimization for Aerial 3D Scanning , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[6]  Gerd Hirzinger,et al.  View Planning for Multi-View Stereo 3D Reconstruction Using an Autonomous Multicopter , 2012, J. Intell. Robotic Syst..

[7]  Liang Li,et al.  Maximum Cohesive Grid of Superpixels for Fast Object Localization , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[8]  Alessio Del Bue,et al.  Directional Visual Descriptors and Multirobot Strategies for Large‐Scale Coverage Problems , 2016, J. Field Robotics.

[9]  C. Ian Connolly,et al.  The determination of next best views , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[10]  Davide Scaramuzza,et al.  A comparison of volumetric information gain metrics for active 3D object reconstruction , 2017, Autonomous Robots.

[11]  Paul Newman,et al.  Surface Edge Explorer (see): Planning Next Best Views Directly from 3D Observations , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[12]  Richard Pito,et al.  A Solution to the Next Best View Problem for Automated Surface Acquisition , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Gaurav S. Sukhatme,et al.  A probabilistic framework for next best view estimation in a cluttered environment , 2014, J. Vis. Commun. Image Represent..

[14]  Joachim Hertzberg,et al.  An autonomous mobile robot with a 3D laser range finder for 3D exploration and digitalization of indoor environments , 2003, Robotics Auton. Syst..

[15]  Rafael Murrieta-Cid,et al.  Volumetric Next-best-view Planning for 3D Object Reconstruction with Positioning Error , 2014 .

[16]  Gamini Dissanayake,et al.  Efficient neighbourhood-based information gain approach for exploration of complex 3D environments , 2013, 2013 IEEE International Conference on Robotics and Automation.

[17]  Karsten Berns,et al.  Dynamic Frontier Based Exploration with a Mobile Indoor Robot , 2010, ISR/ROBOTIK.

[18]  Garrison W. Cottrell,et al.  Belief tree search for active object recognition , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[19]  G. Roth,et al.  View planning for automated three-dimensional object reconstruction and inspection , 2003, CSUR.

[20]  Ashish Kapoor,et al.  Learn-to-Score: Efficient 3D Scene Exploration by Predicting View Utility , 2018, ECCV.

[21]  Thomas A. Funkhouser,et al.  Semantic Scene Completion from a Single Depth Image , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Feng Lin,et al.  A Two-Stage Optimized Next-View Planning Framework for 3-D Unknown Environment Exploration, and Structural Reconstruction , 2017, IEEE Robotics and Automation Letters.

[23]  Cyrill Stachniss,et al.  Effective Exploration for MAVs Based on the Expected Information Gain , 2018 .

[24]  Michael Suppa,et al.  Efficient next-best-scan planning for autonomous 3D surface reconstruction of unknown objects , 2013, Journal of Real-Time Image Processing.

[25]  Thomas Deselaers,et al.  Measuring the Objectness of Image Windows , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[26]  Timothy Patten,et al.  Monte Carlo planning for active object classification , 2017, Autonomous Robots.