Grasp Detection Based on Faster Region CNN

The robot grasp technology has recently aroused increasing research interests thanks to its foundation and importance in the field of robotics. Based on the deep learning method, this paper introduces a grasp detection model with the improved model Faster Region Convolutional Neural Network (Faster-RCNN). The orientation of the ground truth box in the grasp detection is random, so the orientation issue is one of the key points in grasp detection and differs from the other object detection researches. To tackle with this problem, this paper applies the five-dimensional parameters to represent the grasp rectangle. The method puts forward the improved Region Proposal Network (RPN) to export the tilted graspable region, including the size, the location, the orientation and the score belongs to the grasp class or non-grasp class. The RPN extracts the candidate proposals via using a more efficient CNN, instead of the inefficient selective search method. In the classification branch, the softmax function works to determine whether the anchor box is foreground or background. The regression of the angle is performed in the regression branch. In addition, the improved Non-Maximum Suppression (NMS) is used to generate the optimal inclined predicted grasp rectangle. To cope with the insufficient data size in the Cornell Grasp Dataset, the data augmentation and transfer learning methods are applied in the training phase. During the test, the results reveal that the detection accuracy of the model proposed in this paper on the dataset is 92.3% in terms of the image-wise splitting and 92.5% with respect to the objective-wise splitting on the Cornel Grasp Dataset, respectively.

[1]  Abhishek Das,et al.  Object Grasping using Convolutional Neural Networks , 2019, 2019 International Conference on Vision Towards Emerging Trends in Communication and Networking (ViTECoN).

[2]  Henrik I. Christensen,et al.  Automatic grasp planning using shape primitives , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[3]  Peter K. Allen,et al.  An SVM learning approach to robotic grasping , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[4]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Siddhartha S. Srinivasa,et al.  Physics-Based Grasp Planning Through Clutter , 2012, Robotics: Science and Systems.

[6]  Andrew Y. Ng,et al.  Text Detection and Character Recognition in Scene Images with Unsupervised Feature Learning , 2011, 2011 International Conference on Document Analysis and Recognition.

[7]  Patricio A. Vela,et al.  Real-World Multiobject, Multigrasp Detection , 2018, IEEE Robotics and Automation Letters.

[8]  Wolfram Burgard,et al.  An evaluation of the RGB-D SLAM system , 2012, 2012 IEEE International Conference on Robotics and Automation.

[9]  Stefan Ulbrich,et al.  OpenGRASP: A Toolkit for Robot Grasping Simulation , 2010, SIMPAR.

[10]  Vijay Kumar,et al.  Robotic grasping and contact: a review , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).

[11]  Peter K. Allen,et al.  Graspit! A versatile simulator for robotic grasping , 2004, IEEE Robotics & Automation Magazine.

[12]  Dieter Fox,et al.  A large-scale hierarchical multi-view RGB-D object dataset , 2011, 2011 IEEE International Conference on Robotics and Automation.

[13]  Christopher Kanan,et al.  Robotic grasp detection using deep convolutional neural networks , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[14]  Martin A. Riedmiller,et al.  A learned feature descriptor for object recognition in RGB-D data , 2012, 2012 IEEE International Conference on Robotics and Automation.

[15]  Dieter Fox,et al.  Detection-based object labeling in 3D scenes , 2012, 2012 IEEE International Conference on Robotics and Automation.

[16]  Alberto Rodriguez,et al.  Learning Synergies Between Pushing and Grasping with Self-Supervised Deep Reinforcement Learning , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[17]  Joseph Redmon,et al.  Real-time grasp detection using convolutional neural networks , 2014, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Matei T. Ciocarlie,et al.  The Columbia grasp database , 2009, 2009 IEEE International Conference on Robotics and Automation.

[19]  Dieter Fox,et al.  RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments , 2012, Int. J. Robotics Res..

[20]  Honglak Lee,et al.  Deep learning for detecting robotic grasps , 2013, Int. J. Robotics Res..

[21]  Xinyu Liu,et al.  Dex-Net 3.0: Computing Robust Robot Suction Grasp Targets in Point Clouds using a New Analytic Model and Deep Learning , 2017, ArXiv.

[22]  Ashutosh Saxena,et al.  Efficient grasping from RGBD images: Learning using a new rectangle representation , 2011, 2011 IEEE International Conference on Robotics and Automation.