Position-based monocular visual servoing of an unknown target using online self-supervised learning

Visual servoing, i.e. control with visual information, is a valuable capability in many robotic applications. In particular, position based visual servoing (PBVS) estimates position information from the observed image to generate visual servo control. However, the estimation of the position of an unknown target using monocular images is still difficult due to the complexity of the image information. For the target estimation problem, we propose to integrate three complementary techniques for monocular visual servoing. First, to estimate the probability of a target’s existence, the learning model with spatial features from convolution neural network is proposed. Second, the extended Kalman filter based on epipolar geometry estimates the 3D position of the target; moreover, from this 3D position, the perception model is trained online by self-generated virtual ground-truth. Finally, visual servo control is generated, and the resulting movement helps to construct epipolar geometry. Finally. the experimental validation is performed in a challenging setting involving occlusion and target’s shape change.

[1]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[2]  Mohammed Marey,et al.  A Kalman-Filter-Based Method for Pose Estimation in Visual Servoing , 2010, IEEE Transactions on Robotics.

[3]  Xinwu Liang,et al.  Visual Servoing of Soft Robot Manipulator in Constrained Environments With an Adaptive Controller , 2017, IEEE/ASME Transactions on Mechatronics.

[4]  Gangqi Dong,et al.  Position-based visual servo control of autonomous robotic manipulators , 2015 .

[5]  Nitish Srivastava Unsupervised Learning of Visual Representations using Videos , 2015 .

[6]  Éric Marchand,et al.  Histograms-Based Visual Servoing , 2017, IEEE Robotics and Automation Letters.

[7]  Christophe Collewet,et al.  Photometric Visual Servoing , 2011, IEEE Transactions on Robotics.

[8]  Peter I. Corke,et al.  Training Deep Neural Networks for Visual Servoing , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[9]  Ying Wang,et al.  Mobile Robot Regulation with Position Based Visual Servoing , 2018, 2018 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA).

[10]  Éric Marchand,et al.  Visual servoing through mirror reflection , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[11]  Jiang Wang,et al.  3D relative position and orientation estimation using Kalman filter for robot control , 1992, Proceedings 1992 IEEE International Conference on Robotics and Automation.

[12]  Patrick Rives,et al.  A new approach to visual servoing in robotics , 1992, IEEE Trans. Robotics Autom..

[13]  Geraldo F. Silveira,et al.  Direct Visual Servoing: Vision-Based Estimation and Control Using Only Nonmetric Information , 2012, IEEE Transactions on Robotics.

[14]  François Chaumette,et al.  Visual servo control. I. Basic approaches , 2006, IEEE Robotics & Automation Magazine.

[15]  Xuebo Zhang,et al.  Visual Servoing of Nonholonomic Mobile Robots With Uncalibrated Camera-to-Robot Parameters , 2017, IEEE Transactions on Industrial Electronics.

[16]  Xiaogang Wang,et al.  Visual Tracking with Fully Convolutional Networks , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[17]  Sergey Levine,et al.  Learning Visual Servoing with Deep Features and Fitted Q-Iteration , 2017, ICLR.

[18]  Ali Farhadi,et al.  Re$^3$: Re al-Time Recurrent Regression Networks for Visual Tracking of Generic Objects , 2017, IEEE Robotics and Automation Letters.

[19]  De Xu,et al.  Partially Decoupled Image-Based Visual Servoing Using Different Sensitive Features , 2017, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[20]  Luca Bertinetto,et al.  Staple: Complementary Learners for Real-Time Tracking , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[21]  Xinwu Liang,et al.  Image-Based Visual Servoing of a Quadrotor Using Virtual Camera Approach , 2017, IEEE/ASME Transactions on Mechatronics.

[22]  Yuval Tassa,et al.  Control-limited differential dynamic programming , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[23]  Bohyung Han,et al.  Learning Multi-domain Convolutional Neural Networks for Visual Tracking , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Maurizio Ficocelli,et al.  Adaptive filtering for pose estimation in visual servoing , 2001, Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems. Expanding the Societal Role of Robotics in the the Next Millennium (Cat. No.01CH37180).

[25]  Rui Caseiro,et al.  High-Speed Tracking with Kernelized Correlation Filters , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[26]  Sergey Levine,et al.  Deep visual foresight for planning robot motion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[27]  William J. Wilson,et al.  Relative end-effector control using Cartesian position based visual servoing , 1996, IEEE Trans. Robotics Autom..

[28]  K. Madhava Krishna,et al.  Exploring convolutional networks for end-to-end visual servoing , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[29]  Éric Marchand,et al.  Mutual Information-Based Visual Servoing , 2011, IEEE Transactions on Robotics.