A New Approach Based on Two-stream CNNs for Novel Objects Grasping in Clutter

Recently, many researches focus on learning to grasp novel objects, which is an important but still unsolved issue especially for service robots. While some approaches perform well in some cases, they need human labeling and can hardly be used in clutter with a high precision. In this paper, we apply a deep learning approach to solve the problem about grasping novel objects in clutter. We focus on two-fingered parallel-jawed grasping with RGBD camera. Firstly, we propose a ‘grasp circle’ method to find more potential grasps in each sampling point with less cost, which is parameterized by the size of the gripper. Considering the challenge of collecting large amounts of training data, we collect training data directly from cluttered scene with no manual labeling. Then we need to extract effective features from RGB and depth data, for which we propose a bimodal representation and use two-stream convolution neural networks (CNNs) to handle the processed inputs. Finally the experiment shows that compared to some existing popular methods, our method gets higher success rate of grasping for the original RGB-D cluttered scene.

[1]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[2]  Van-Duc Nguyen,et al.  Constructing Force- Closure Grasps , 1988, Int. J. Robotics Res..

[3]  Guanglin Li,et al.  Development of Sensory-Motor Fusion-Based Manipulation and Grasping Control for a Robotic Hand-Eye System , 2017, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[4]  Abhinav Gupta,et al.  Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Kate Saenko,et al.  High precision grasp pose detection in dense clutter , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[6]  Kate Saenko,et al.  Grasp Pose Detection in Point Clouds , 2017, Int. J. Robotics Res..

[7]  Martín Abadi,et al.  TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.

[8]  Joseph Redmon,et al.  Real-time grasp detection using convolutional neural networks , 2014, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[9]  Alexander Herzog,et al.  Template-based learning of grasp selection , 2012, 2012 IEEE International Conference on Robotics and Automation.

[10]  Danica Kragic,et al.  Data-Driven Grasp Synthesis—A Survey , 2013, IEEE Transactions on Robotics.

[11]  Kosei Kitagaki,et al.  Stability of soft-finger grasp under gravity , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[12]  Jason Jianjun Gu,et al.  Robotic grasp detection using extreme learning machine , 2015, 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[13]  Wolfram Burgard,et al.  Multimodal deep learning for robust RGB-D object recognition , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[14]  Ashutosh Saxena,et al.  Efficient grasping from RGBD images: Learning using a new rectangle representation , 2011, 2011 IEEE International Conference on Robotics and Automation.

[15]  Jeannette Bohg,et al.  Leveraging big data for grasp planning , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Nicolas Chopin,et al.  Fast simulation of truncated Gaussian distributions , 2011, Stat. Comput..

[17]  Olaf Kähler,et al.  Very High Frame Rate Volumetric Integration of Depth Images on Mobile Devices , 2015, IEEE Transactions on Visualization and Computer Graphics.

[18]  Devendra P. Garg,et al.  Grasp Planning for a Two-Fingered Parallel-Jawed Gripper , 2006 .

[19]  Markus Vincze,et al.  Learning grasps with topographic features , 2015, Int. J. Robotics Res..

[20]  Sergey Levine,et al.  Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection , 2016, Int. J. Robotics Res..

[21]  Robert Platt,et al.  Using Geometry to Detect Grasp Poses in 3D Point Clouds , 2015, ISRR.

[22]  Markus Vincze,et al.  Empty the basket - a shape based learning approach for grasping piles of unknown objects , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[23]  Honglak Lee,et al.  Deep learning for detecting robotic grasps , 2013, Int. J. Robotics Res..

[24]  Peter K. Allen,et al.  Graspit! A versatile simulator for robotic grasping , 2004, IEEE Robotics & Automation Magazine.

[25]  Tamim Asfour,et al.  Planning grasps for robotic hands using a novel object representation based on the medial axis transform , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[26]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[27]  Quoc V. Le,et al.  Learning to grasp objects with multiple contact points , 2010, 2010 IEEE International Conference on Robotics and Automation.

[28]  Tian Gao,et al.  Efficient Markov Blanket Discovery and Its Application , 2017, IEEE Transactions on Cybernetics.

[29]  Takeo Kanade,et al.  Automated Construction of Robotic Manipulation Programs , 2010 .

[30]  Ashutosh Saxena,et al.  Learning to Grasp Novel Objects Using Vision , 2006, ISER.

[31]  Xinyu Liu,et al.  Dex-Net 2.0: Deep Learning to Plan Robust Grasps with Synthetic Point Clouds and Analytic Grasp Metrics , 2017, Robotics: Science and Systems.