Visual Servo Tracking of a Wheeled Mobile Robot Utilizing Both Feature and Depth Information

This paper presents a novel visual servo tracking method for a wheeled mobile robot based on target features and depth of scenery. The robot is equipped with an RGBD sensor to provide both color and depth information of the scene. The proposed method comprises two phases. The first phase utilizes feature matching to ensure that the target moves into the field of vision such that the robot may enter the second phase. The second phase implements visual servo aiming at object tacking accuracy. The depth information is employed in the proposed approach, which reduces the computational complexity required as in conventional approaches based on feature extraction. In addition, the second phase has built-in mechanism for detecting and exiting local minimum when the robot speed converges. The method is also capable of sequentially tracking multiple objects. Experimental results demonstrate the feasibility and effectiveness of the proposed two-phase method.

[1]  Alexander Wong,et al.  Difference of Circles Feature Detector , 2012, 2012 Ninth Conference on Computer and Robot Vision.

[2]  Éric Marchand,et al.  A Dense and Direct Approach to Visual Servoing Using Depth Maps , 2014, IEEE Transactions on Robotics.

[3]  Robert E. Mahony,et al.  Visual servoing of an under-actuated dynamic rigid-body system: an image-based approach , 2002, IEEE Trans. Robotics Autom..

[4]  Lee E. Weiss,et al.  Dynamic sensor-based control of robots with visual feedback , 1987, IEEE Journal on Robotics and Automation.

[5]  Philippe Martinet,et al.  Vision Based Control Law using 3D Visual Features , 1997 .

[6]  Luc Van Gool,et al.  SURF: Speeded Up Robust Features , 2006, ECCV.

[7]  François Chaumette,et al.  2 1/2 D visual servoing: a possible solution to improve image-based and position-based visual servoings , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).

[8]  Bill Triggs,et al.  Histograms of oriented gradients for human detection , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[9]  William J. Wilson,et al.  Relative end-effector control using Cartesian position based visual servoing , 1996, IEEE Trans. Robotics Autom..

[10]  Kurt Konolige,et al.  CenSurE: Center Surround Extremas for Realtime Feature Detection and Matching , 2008, ECCV.

[11]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[12]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[13]  John M. Hollerbach,et al.  Autonomous Robot Calibration for Hand-Eye Coordination , 1991, Int. J. Robotics Res..

[14]  François Chaumette,et al.  Visual servo control. I. Basic approaches , 2006, IEEE Robotics & Automation Magazine.