TossingBot: Learning to Throw Arbitrary Objects With Residual Physics

We investigate whether a robot arm can learn to pick and throw arbitrary objects into selected boxes quickly and accurately. Throwing has the potential to increase the physical reachability and picking speed of a robot arm. However, precisely throwing arbitrary objects in unstructured settings presents many challenges: from acquiring reliable pre-throw conditions (e.g. initial pose of object in manipulator) to handling varying object-centric properties (e.g. mass distribution, friction, shape) and dynamics (e.g. aerodynamics). In this work, we propose an end-to-end formulation that jointly learns to infer control parameters for grasping and throwing motion primitives from visual observations (images of arbitrary objects in a bin) through trial and error. Within this formulation, we investigate the synergies between grasping and throwing (i.e., learning grasps that enable more accurate throws) and between simulation and deep learning (i.e., using deep networks to predict residuals on top of control parameters predicted by a physics simulator). The resulting system, TossingBot, is able to grasp and throw arbitrary objects into boxes located outside its maximum reach range at 500+ mean picks per hour (600+ grasps per hour with 85% throwing accuracy); and generalizes to new objects and target locations. Videos are available at this https URL

[1]  Leslie Pack Kaelbling,et al.  Augmenting Physical Simulators with Stochastic Neural Networks: Case Study of Planar Pushing and Bouncing , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[2]  Jiajun Wu,et al.  DensePhysNet: Learning Dense Physical Object Representations via Multi-step Dynamic Interactions , 2019, Robotics: Science and Systems.

[3]  Andy Zeng,et al.  Learning to See before Learning to Act: Visual Pre-training for Manipulation , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[4]  Ian Taylor,et al.  Robotic pick-and-place of novel objects in clutter with multi-affordance grasping and cross-domain image matching , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Masatoshi Ishikawa,et al.  High-speed throwing motion based on kinetic chain approach , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[6]  Gaurav S. Sukhatme,et al.  Learning task error models for manipulation , 2013, 2013 IEEE International Conference on Robotics and Automation.

[7]  M.T. Mason,et al.  Dynamic manipulation , 1993, Proceedings of 1993 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '93).

[8]  Roberto Cipolla,et al.  SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Andy Zeng,et al.  Learning Visual Affordances for Robotic Manipulation , 2019 .

[10]  Leslie Pack Kaelbling,et al.  Residual Policy Learning , 2018, ArXiv.

[11]  Ross B. Girshick,et al.  Fast R-CNN , 2015, 1504.08083.

[12]  Pieter Abbeel,et al.  Using inaccurate models in reinforcement learning , 2006, ICML.

[13]  Jan Peters,et al.  Reinforcement Learning to Adjust Robot Movements to New Situations , 2010, IJCAI.

[14]  Matthias Nießner,et al.  3DMatch: Learning Local Geometric Descriptors from RGB-D Reconstructions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Alberto Rodriguez,et al.  Optimal shape and motion planning for dynamic planar manipulation , 2018, Auton. Robots.

[16]  Marc Levoy,et al.  A volumetric method for building complex models from range images , 1996, SIGGRAPH.

[17]  Takeo Kanade,et al.  Automated Construction of Robotic Manipulation Programs , 2010 .

[18]  C. Karen Liu,et al.  Data-Augmented Contact Model for Rigid Body Simulation , 2018, L4DC.

[19]  Peter I. Corke,et al.  Cartman: The Low-Cost Cartesian Manipulator that Won the Amazon Robotics Challenge , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[20]  Shuran Song,et al.  Clear Grasp: 3D Shape Estimation of Transparent Objects for Manipulation , 2019, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[21]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Ken Goldberg,et al.  Learning ambidextrous robot grasping policies , 2019, Science Robotics.

[23]  Sergey Levine,et al.  Residual Reinforcement Learning for Robot Control , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[24]  Danica Kragic,et al.  Deep predictive policy training using reinforcement learning , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[25]  Alberto Rodriguez,et al.  Learning Synergies Between Pushing and Grasping with Self-Supervised Deep Reinforcement Learning , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[26]  Kevin M. Lynch,et al.  Dynamic Nonprehensile Manipulation: Controllability, Planning, and Experiments , 1999, Int. J. Robotics Res..

[27]  Gregory Dudek,et al.  Adapting learned robotics behaviours through policy adjustment , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[28]  Andrew W. Fitzgibbon,et al.  KinectFusion: Real-time dense surface mapping and tracking , 2011, 2011 10th IEEE International Symposium on Mixed and Augmented Reality.

[29]  David J. Reinkensmeyer,et al.  Task-level robot learning , 1988, Proceedings. 1988 IEEE International Conference on Robotics and Automation.

[30]  Siddhartha S. Srinivasa,et al.  Extrinsic dexterity: In-hand manipulation with external forces , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[31]  Jwu-Sheng Hu,et al.  A ball-throwing robot with visual feedback , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[32]  Yukinori Kobayashi,et al.  Motion Control of a Ball Throwing Robot with a Flexible Robotic Arm , 2013 .

[33]  Stefan Schaal,et al.  Combining learned and analytical models for predicting action effects from sensory data , 2017, Int. J. Robotics Res..

[34]  Tom Schaul,et al.  Prioritized Experience Replay , 2015, ICLR.

[35]  Xinyu Liu,et al.  Dex-Net 2.0: Deep Learning to Plan Robust Grasps with Synthetic Point Clouds and Analytic Grasp Metrics , 2017, Robotics: Science and Systems.

[36]  Nima Fazeli,et al.  Learning Data-Efficient Rigid-Body Contact Models: Case Study of Planar Impact , 2017, CoRL.

[37]  Tsukasa Ogasawara,et al.  1-DOF dynamic pitching robot that independently controls velocity, Angular velocity, and direction of a ball: Contact models and motion planning , 2009, 2009 IEEE International Conference on Robotics and Automation.