Recovering Scale in Relative Pose and Target Model Estimation Using Monocular Vision

A combined relative pose and target object model estimation framework using a monocular camera as the primary feedback sensor has been designed and validated in a simulated robotic environment. The monocular camera is mounted on the end-effector of a robot manipulator and measures the image plane coordinates of a set of point features on a target workpiece object. Using this information, the relative position and orientation, as well as the geometry, of the target object are recovered recursively by a Kalman filter process. The Kalman filter facilitates the fusion of supplemental measurements from range sensors, with those gathered with the camera. This process allows the estimated system state to be accurate and recover the proper environment scale. Current approaches in the research areas of visual servoing control and mobile robotics are studied in the case where the target object feature point geometry is well-known prior to the beginning of the estimation. In this case, only the relative pose of target object frames is estimated over a sequence of frames from a single monocular camera. An observability analysis was carried out to identify the physical configurations of camera and target object for which the relative pose cannot be recovered by measuring only the camera image plane coordinates of the object point features. A popular extension to this is to concurrently estimate the target object model concurrently with the relative pose of the camera frame, a process known as Simultaneous Localization and Mapping (SLAM). The recursive framework was augmented to facilitate this larger estimation problem. The scale of the recovered solution is ambiguous using measurements from a single camera. A second observability analysis highlights more configurations for which the relative pose and target object model are unrecoverable from camera measurements alone. Instead, measurements which contain the global scale are required to obtain an accurate solution. A set of additional sensors are detailed, including range finders and additional cameras. Measurement models for each are given, which facilitate the fusion of this supplemental data with the original monocular camera image measurements. A complete framework is then derived to combine a set of such sensor measurements to recover an accurate relative pose and target object model estimate. This proposed framework is tested in a simulation environment with a virtual robot manipulator tracking a target object workpiece through a relative trajectory. All of the detailed estimation schemes are executed: the single monocular camera

[1]  François Chaumette,et al.  Visual servo control. II. Advanced approaches [Tutorial] , 2007, IEEE Robotics & Automation Magazine.

[2]  Mark W. Spong,et al.  Robot dynamics and control , 1989 .

[3]  Martial Hebert,et al.  Experimental Comparison of Techniques for Localization and Mapping Using a Bearing-Only Sensor , 2000, ISER.

[4]  C. V. Jawahar,et al.  Target Model Estimation using Particle Filters for Visual Servoing , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[5]  Philippe Martinet,et al.  Position based visual servoing using a non-linear approach , 1999, Proceedings 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289).

[6]  F. Markley,et al.  Attitude Estimation Using Modified Rodrigues Parameters , 1996 .

[7]  Tom Drummond,et al.  Scalable Monocular SLAM , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[8]  Yolanda González Cid,et al.  Real-time 3d SLAM with wide-angle vision , 2004 .

[9]  Stefano Soatto Observability/identifiability of rigid motion under perspective projection , 1994, Proceedings of 1994 33rd IEEE Conference on Decision and Control.

[10]  Takeo Kanade,et al.  Development of a Video-Rate Stereo Machine , 1997 .

[11]  A. Krener,et al.  Nonlinear controllability and observability , 1977 .

[12]  William J. Wilson,et al.  Relative end-effector control using Cartesian position based visual servoing , 1996, IEEE Trans. Robotics Autom..

[13]  Tsukasa Ogasawara,et al.  Fast self-localization method for mobile robots using multiple omnidirectional vision sensors , 2003, Machine Vision and Applications.

[14]  Kurt Konolige,et al.  Frame-Frame Matching for Realtime Consistent Visual Mapping , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[15]  Eric T. Baumgartner,et al.  An autonomous vision-based mobile robot , 1994, IEEE Trans. Autom. Control..

[16]  Hyun Myung,et al.  Structured light 2D range finder for simultaneous localization and map-building (SLAM) in home environments , 2004, Micro-Nanomechatronics and Human Science, 2004 and The Fourth Symposium Micro-Nanomechatronics for Information-Based Society, 2004..

[17]  Wolfram Burgard,et al.  Probabilistic Robotics (Intelligent Robotics and Autonomous Agents) , 2005 .

[18]  E. J. Haug,et al.  Computer aided kinematics and dynamics of mechanical systems. Vol. 1: basic methods , 1989 .

[19]  François Chaumette,et al.  Visual servo control. I. Basic approaches , 2006, IEEE Robotics & Automation Magazine.

[20]  Carol Hulls,et al.  Integration of camera and range sensors for 3D pose estimation in robot visual servoing , 1998, Other Conferences.

[21]  Javier Civera,et al.  Inverse Depth to Depth Conversion for Monocular SLAM , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[22]  Ze-Nian Li,et al.  A survey of motion-parallax-based 3-D reconstruction algorithms , 2004, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[23]  Lee E. Weiss,et al.  Dynamic sensor-based control of robots with visual feedback , 1987, IEEE Journal on Robotics and Automation.

[24]  Seth Hutchinson,et al.  Visual Servo Control Part I: Basic Approaches , 2006 .

[25]  I. Bar-Itzhack,et al.  Observability analysis of piece-wise constant systems. I. Theory , 1992 .

[26]  Juan Andrade-Cetto,et al.  The effects of partial observability in SLAM , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[27]  Patrick Rives,et al.  Singularities in the determination of the situation of a robot effector from the perspective view of 3 points , 1993 .

[28]  Andrew J. Davison,et al.  Real-time simultaneous localisation and mapping with a single camera , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[29]  Javier Civera,et al.  Dimensionless Monocular SLAM , 2007, IbPRIA.

[30]  Javier Civera,et al.  Inverse Depth Parametrization for Monocular SLAM , 2008, IEEE Transactions on Robotics.

[31]  José A. Castellanos,et al.  Simultaneous map building and localization for mobile robots: a multisensor fusion approach , 1998, Proceedings. 1998 IEEE International Conference on Robotics and Automation (Cat. No.98CH36146).

[32]  Philippe Bonnifait,et al.  Design and experimental validation of an odometric and goniometric localization system for outdoor robot vehicles , 1998, IEEE Trans. Robotics Autom..

[33]  Avinash C. Kak,et al.  Vision for Mobile Robot Navigation: A Survey , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[34]  Ki-Sang Hong,et al.  Vision-based simultaneous localization and mapping with two cameras , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[35]  Teresa A. Vidal-Calleja,et al.  On the Observability of Bearing-only SLAM , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[36]  Larry S. Davis,et al.  Model-based object pose in 25 lines of code , 1992, International Journal of Computer Vision.

[37]  Sebastian Thrun,et al.  FastSLAM 2.0: An Improved Particle Filtering Algorithm for Simultaneous Localization and Mapping that Provably Converges , 2003, IJCAI.

[38]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[39]  T. Başar,et al.  A New Approach to Linear Filtering and Prediction Problems , 2001 .

[40]  Avinash C. Kak,et al.  A New Kalman-Filter-Based Framework for Fast and Accurate Visual Tracking of Rigid Objects , 2008, IEEE Transactions on Robotics.

[41]  David H. Eberly,et al.  3D game engine design - a practical approach to real-time computer graphics , 2000 .

[42]  E. Malis,et al.  2 1/2 D Visual Servoing , 1999 .

[43]  W. Wilson,et al.  Decoupled EKF for simultaneous target model and relative pose estimation using feature points , 2005, Proceedings of 2005 IEEE Conference on Control Applications, 2005. CCA 2005..

[44]  Avinash C. Kak,et al.  Vision-based navigation of mobile robot with obstacle avoidance by single camera vision and ultrasonic sensing , 1997, Proceedings of the 1997 IEEE/RSJ International Conference on Intelligent Robot and Systems. Innovative Robotics for Real-World Applications. IROS '97.

[45]  Frank Dellaert,et al.  Visual SLAM with a Multi-Camera Rig , 2006 .

[46]  Wang Wei,et al.  Mobile Robot Indoor Simultaneous Localization and Mapping Using Laser Range Finder and Monocular Vision , 2005 .

[47]  Katsuhiko Ogata,et al.  Modern Control Engineering , 1970 .

[48]  Antonio Bicchi,et al.  Visual Servoing in the Large , 2009, Int. J. Robotics Res..

[49]  Matteo Matteucci,et al.  Monocular SLAM with Inverse Scaling Parametrization , 2008, BMVC.

[50]  Hugh Durrant-Whyte,et al.  Simultaneous localization and mapping (SLAM): part II , 2006 .

[51]  Teresa A. Vidal-Calleja,et al.  Fusing Monocular Information in Multicamera SLAM , 2008, IEEE Transactions on Robotics.

[52]  Hugh F. Durrant-Whyte,et al.  A solution to the simultaneous localization and map building (SLAM) problem , 2001, IEEE Trans. Robotics Autom..

[53]  George K. I. Mann,et al.  Integrated laser-camera sensor for the detection and localization of landmarks for robotic applications , 2008, 2008 IEEE International Conference on Robotics and Automation.

[54]  S. Hutchinson,et al.  Visual Servo Control Part II : Advanced Approaches , 2007 .