Vision-based Indoor Positioning Method By Joint Using 2D Images and 3D Point Cloud Map

2D image-based and 3D structure-based are the most two popular vision-based indoor positioning methods. Though 2D method has better efficiency for image retrieval, the number of stored images is usually too many to achieve fast 2D-2D matching, and it cannot obtain high positioning accuracy either. In contrast, 3D method has smaller database and can achieve much higher positioning accuracy with the help of 3D information. However, it needs to construct 3D point cloud map of the positioning environment and spend too much time on 2D-3D matching. Therefore, in order to improve the positioning accuracy and the retrieval efficiency, we propose a vision-based indoor positioning method by joint use 2D images and 3D point cloud map in this paper. In offline stage, we select key images into the offline database based on the 3D point cloud map that is built by RGB-D simultaneous localization and mapping (SLAM), and establish the mapping relation between 2D image local feature points and 3D points. In online stage, we utilize the rough-fine matching strategy to obtain 2D-3D matching pairs between local feature points of the user query image and 3D points in the 3D point cloud map. The user position is finally estimated by the efficient perspective-n-point (EPnP) algorithm. Performance analysis and simulation results indicate that our proposed method can not only achieve high positioning accuracy but also speed up the matching process.

[1]  Michael Greenspan,et al.  Visual indoor positioning with a single camera using PnP , 2015, 2015 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[2]  Robert Piché,et al.  A Survey of Selected Indoor Positioning Methods for Smartphones , 2017, IEEE Communications Surveys & Tutorials.

[3]  David Nistér,et al.  An efficient solution to the five-point relative pose problem , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Yiannis Andreopoulos,et al.  Voronoi-Based Compact Image Descriptors: Efficient Region-of-Interest Retrieval With VLAD and Deep-Learning-Based Descriptors , 2016, IEEE Transactions on Multimedia.

[5]  Christopher Hunt,et al.  Notes on the OpenSURF Library , 2009 .

[6]  Michel Antunes,et al.  Piecewise-Planar StereoScan: Sequential Structure and Motion Using Plane Primitives , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  V. Lepetit,et al.  EPnP: An Accurate O(n) Solution to the PnP Problem , 2009, International Journal of Computer Vision.

[8]  Jan-Michael Frahm,et al.  From structure-from-motion point clouds to fast location recognition , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.