Sensor fusion for fiducial tags: Highly robust pose estimation from single frame RGBD

Although there is an abundance of planar fiducial-marker systems proposed for augmented reality and computer-vision purposes, using them to estimate the pose accurately in robotic applications where collected data are noisy remains a challenge. This is inherently a difficult problem because these fiducial marker systems work solely within the RGB image space and the resolution of cameras on robots is often constrained. As a result, small noise in the image would cause the tag's detection process to produce large pose estimation errors. This paper describes an algorithm that improves the pose estimation accuracy of square fiducial markers in difficult scenes by fusing information from RGB and depth sensors. The algorithm retains the high detection rate and low false positive rate characteristics of fiducial systems while making them much more robust to size, lighting and sensory noise for pose estimation. The improvements make the fiducial tags suitable for robotic tasks requiring high pose accuracy in the real world environment.

[1]  Francisco José Madrid-Cuevas,et al.  Automatic generation and detection of highly reliable fiducial markers under occlusion , 2014, Pattern Recognit..

[2]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[3]  Robert B. Fisher,et al.  Estimating 3-D rigid body transformations: a comparison of four major algorithms , 1997, Machine Vision and Applications.

[4]  Paul J. Besl,et al.  Method for registration of 3-D shapes , 1992, Other Conferences.

[5]  Farrokh Janabi-Sharifi,et al.  A Robust Vision-Based Sensor Fusion Approach for Real-Time Pose Estimation , 2014, IEEE Transactions on Cybernetics.

[6]  Eric Foxlin,et al.  Circular data matrix fiducial system and robust image processing for a wearable vision-inertial self-tracker , 2002, Proceedings. International Symposium on Mixed and Augmented Reality.

[7]  Andreas Birk,et al.  Uncertainty analysis for optimum plane extraction from noisy 3D range-sensor point-clouds , 2010, Intell. Serv. Robotics.

[8]  Shahram Izadi,et al.  Modeling Kinect Sensor Noise for Improved 3D Reconstruction and Tracking , 2012, 2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission.

[9]  Andrea Torsello,et al.  RUNE-Tag: A high accuracy fiducial marker with strong occlusion resilience , 2011, CVPR 2011.

[10]  Edwin Olson,et al.  AprilTag 2: Efficient and robust fiducial detection , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[11]  Radu Horaud,et al.  An analytic solution for the perspective 4-point problem , 1989, Proceedings CVPR '89: IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[12]  Robert M. Haralick,et al.  Review and analysis of solutions of the three point perspective pose estimation problem , 1994, International Journal of Computer Vision.

[13]  Daniel Grest,et al.  A Comparison of Iterative 2D-3D Pose Estimation Methods for Real-Time Applications , 2009, SCIA.

[14]  A. Aydin Alatan,et al.  RGBD data based pose estimation: Why sensor fusion? , 2015, 2015 18th International Conference on Information Fusion (Fusion).

[15]  Edwin Olson,et al.  AprilTag: A robust and flexible visual fiducial system , 2011, 2011 IEEE International Conference on Robotics and Automation.

[16]  胡占义,et al.  A General Sufficient Condition of Four Positive Solutions of the P3P Problem , 2005 .

[17]  Alastair R. Beresford,et al.  Analysing fundamental properties of marker-based vision system designs , 2006, Pervasive Mob. Comput..

[18]  Larry S. Davis,et al.  Exact and Approximate Solutions of the Perspective-Three-Point Problem , 1992, IEEE Trans. Pattern Anal. Mach. Intell..