DVGG: Deep Variational Grasp Generation for Dextrous Manipulation

Grasping with anthropomorphic robotic hands involves much more hand-object interactions compared to parallel-jaw grippers. Modeling hand-object interactions is essential to the study of multi-finger hand dextrous manipulation. This work presents DVGG, an efficient grasp generation network that takes single-view observation as input and predicts high-quality grasp configurations for unknown objects. In general, our generative model consists of three components: 1) Point cloud completion for the target object based on the partial observation; 2) Diverse sets of grasps generation given the complete point cloud; 3) Iterative grasp pose refinement for physically plausible grasp optimization. To train our model, we build a large-scale grasping dataset that contains about 300 common object models with 1.5 M annotated grasps in simulation. Experiments in simulation show that our model can predict robust grasp poses with a wide variety and high success rate. Real robot platform experiments demonstrate that the model trained on our dataset performs well in the real world. Remarkably, our method achieves a grasp success rate of 70.7% for novel objects in the real robot platform, which is a significant improvement over the baseline methods.

[1]  Peng Wang,et al.  GPR: Grasp Pose Refinement Network for Cluttered Scenes , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Shuai Yi,et al.  Variational Relational Point Completion Network , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Xiaolong Wang,et al.  Hand-Object Contact Consistency Reasoning for Human Grasps Generation , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[4]  Ville Kyrki,et al.  DDGC: Generative Deep Dexterous Grasping in Clutter , 2021, IEEE Robotics and Automation Letters.

[5]  Cewu Lu,et al.  RGB Matters: Learning 7-DoF Grasp Poses on Monocular RGBD Images , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[6]  Francesc Moreno-Noguer,et al.  Multi-FinGAN: Generative Coarse-To-Fine Sampling of Multi-Finger Grasps , 2020, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Dimitrios Tzionas,et al.  GRAB: A Dataset of Whole-Body Human Grasping of Objects , 2020, ECCV.

[8]  Charles C. Kemp,et al.  ContactPose: A Dataset of Grasps with Object Contact and Hand Pose , 2020, ECCV.

[9]  Tucker Hermans,et al.  Multi-Fingered Active Grasp Learning , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[10]  Francesc Moreno-Noguer,et al.  GanHand: Predicting Human Grasp Affordances in Multi-Object Scenes , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  Cewu Lu,et al.  GraspNet-1Billion: A Large-Scale Benchmark for General Object Grasping , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Peter Corke,et al.  EGAD! An Evolved Grasping Analysis Dataset for Diversity and Reproducibility in Robotic Manipulation , 2020, IEEE Robotics and Automation Letters.

[13]  Feng Xu,et al.  Generative Attention Learning: a “GenerAL” framework for high-performance multi-fingered grasping in clutter , 2020, Autonomous Robots.

[14]  Zherong Pan,et al.  Deep Differentiable Grasp Planner for High-DOF Grippers , 2020, Robotics: Science and Systems.

[15]  Jianlan Luo,et al.  UniGrasp: Learning a Unified Model to Grasp With Multifingered Robotic Hands , 2019, IEEE Robotics and Automation Letters.

[16]  Ales Leonardis,et al.  Deep Dexterous Grasping of Novel Objects from a Single View , 2019, Int. J. Humanoid Robotics.

[17]  Dominik Belter,et al.  Learning better generative models for dexterous, single-view grasping of novel objects , 2019, Int. J. Robotics Res..

[18]  V. Lepetit,et al.  HOnnotate: A Method for 3D Annotation of Hand and Object Poses , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Dieter Fox,et al.  6-DOF GraspNet: Variational Grasp Generation for Object Manipulation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[20]  Michael J. Black,et al.  Learning Joint Reconstruction of Hands and Manipulated Objects , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[21]  Dieter Fox,et al.  ContactGrasp: Functional Multi-finger Grasp Synthesis from Contact , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[22]  Ville Kyrki,et al.  Robust Grasp Planning Over Uncertain Shape Completions , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[23]  Zherong Pan,et al.  Generating Grasp Poses for a High-DOF Gripper Using Neural Networks , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[24]  Fuchun Sun,et al.  PointNetGPD: Detecting Grasp Configurations from Point Sets , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[25]  Shanxin Yuan,et al.  First-Person Hand Action Benchmark with RGB-D Videos and 3D Hand Pose Annotations , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[26]  Xinyu Liu,et al.  Dex-Net 2.0: Deep Learning to Plan Robust Grasps with Synthetic Point Clouds and Analytic Grasp Metrics , 2017, Robotics: Science and Systems.

[27]  Leonidas J. Guibas,et al.  PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  Christopher Kanan,et al.  Robotic grasp detection using deep convolutional neural networks , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[29]  Chad DeChant,et al.  Shape completion enabled robotic grasping , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[30]  Danica Kragic,et al.  The GRASP Taxonomy of Human Grasp Types , 2016, IEEE Transactions on Human-Machine Systems.

[31]  Peter K. Allen,et al.  Generating multi-fingered robotic grasps via deep learning , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[32]  Honglak Lee,et al.  Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.

[33]  P. Abbeel,et al.  Benchmarking in Manipulation Research: The YCB Object and Model Set and Benchmarking Protocols , 2015, ArXiv.

[34]  Joseph Redmon,et al.  Real-time grasp detection using convolutional neural networks , 2014, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[35]  Yu Zheng,et al.  An Efficient Algorithm for a Grasp Quality Measure , 2013, IEEE Transactions on Robotics.

[36]  Yuval Tassa,et al.  MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[37]  Peter K. Allen,et al.  Graspit! A versatile simulator for robotic grasping , 2004, IEEE Robotics & Automation Magazine.

[38]  Gerd Hirzinger,et al.  A fast and robust grasp planner for arbitrary 3D objects , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).

[39]  John F. Canny,et al.  Planning optimal grasps , 1992, Proceedings 1992 IEEE International Conference on Robotics and Automation.

[40]  Van-Duc Nguyen,et al.  Constructing Force- Closure Grasps , 1988, Int. J. Robotics Res..

[41]  Russ Tedrake,et al.  Synthesis and Optimization of Force Closure Grasps via Sequential Semidefinite Programming , 2015, ISRR.

[42]  P. Allen,et al.  Dexterous Grasping via Eigengrasps : A Low-dimensional Approach to a High-complexity Problem , 2007 .