Human-like Planning for Reaching in Cluttered Environments

Humans, in comparison to robots, are remarkably adept at reaching for objects in cluttered environments. The best existing robot planners are based on random sampling of configuration space- which becomes excessively high-dimensional with large number of objects. Consequently, most planners often fail to efficiently find object manipulation plans in such environments. We addressed this problem by identifying high-level manipulation plans in humans, and transferring these skills to robot planners. We used virtual reality to capture human participants reaching for a target object on a tabletop cluttered with obstacles. From this, we devised a qualitative representation of the task space to abstract the decision making, irrespective of the number of obstacles. Based on this representation, human demonstrations were segmented and used to train decision classifiers. Using these classifiers, our planner produced a list of waypoints in task space. These waypoints provided a high-level plan, which could be transferred to an arbitrary robot model and used to initialise a local trajectory optimiser. We evaluated this approach through testing on unseen human VR data, a physics-based robot simulation, and a real robot (dataset and code are publicly available1). We found that the human-like planner outperformed a state-of-the-art standard trajectory optimisation algorithm, and was able to generate effective strategies for rapid planning- irrespective of the number of obstacles in the environment.

[1]  Yuval Tassa,et al.  DeepMind Control Suite , 2018, ArXiv.

[2]  Kazuhiko Kawamura,et al.  A computational framework for integrating robotic exploration and human demonstration in imitation learning , 2011, 2011 IEEE International Conference on Systems, Man, and Cybernetics.

[3]  Luc De Raedt,et al.  Relational affordances for multiple-object manipulation , 2017, Autonomous Robots.

[4]  Sandra Hirche,et al.  Feedback motion planning and learning from demonstration in physical robotic assistance: differences and synergies , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[5]  Maya Cakmak,et al.  Keyframe-based Learning from Demonstration , 2012, Int. J. Soc. Robotics.

[6]  Byron Boots,et al.  Towards Robust Skill Generalization: Unifying Learning from Demonstration and Motion Planning , 2017, CoRL.

[7]  M. Kubát An Introduction to Machine Learning , 2017, Springer International Publishing.

[8]  Ken Goldberg,et al.  Multi-Task Hierarchical Imitation Learning for Home Automation , 2019, 2019 IEEE 15th International Conference on Automation Science and Engineering (CASE).

[9]  Anthony G. Cohn,et al.  Thinking Inside the Box: A Comprehensive Spatial Representation for Video Analysis , 2012, KR.

[10]  Federico Pecora,et al.  More knowledge on the table: Planning with space, time and resources for robots , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[11]  Ashwin P. Dani,et al.  Learning Contracting Nonlinear Dynamics From Human Demonstration for Robot Motion Planning , 2015, HRI 2015.

[12]  Mark Moll,et al.  Randomized Physics-Based Motion Planning for Grasping in Cluttered and Uncertain Environments , 2017, IEEE Robotics and Automation Letters.

[13]  Mehmet Remzi Dogar,et al.  Combining Coarse and Fine Physics for Manipulation using Parallel-in-Time Integration , 2019, ISRR.

[14]  Yuval Tassa,et al.  MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[15]  Mehmet Remzi Dogar,et al.  Real-Time Online Re-Planning for Grasping Under Clutter and Uncertainty , 2018, 2018 IEEE-RAS 18th International Conference on Humanoid Robots (Humanoids).

[16]  Mehmet R. Dogar,et al.  Pushing Fast and Slow: Task-Adaptive Planning for Non-prehensile Manipulation Under Uncertainty , 2018, WAFR.

[17]  Mark H. Overmars,et al.  A Comparative Study of Probabilistic Roadmap Planners , 2002, WAFR.

[18]  G. Swaminathan Robot Motion Planning , 2006 .

[19]  Ken Goldberg,et al.  Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation , 2017, ICRA.

[20]  Emilio Frazzoli,et al.  Sampling-based algorithms for optimal motion planning , 2011, Int. J. Robotics Res..

[21]  Siddhartha S. Srinivasa,et al.  Nonprehensile whole arm rearrangement planning on physics manifolds , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[22]  Marco Pavone,et al.  Learning Sampling Distributions for Robot Motion Planning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[23]  Lydia E. Kavraki,et al.  A Sampling-Based Tree Planner for Systems With Complex Dynamics , 2012, IEEE Transactions on Robotics.

[24]  Faisal Mushtaq,et al.  Studying human behavior with virtual reality: The Unity Experiment Framework , 2019, Behavior research methods.

[25]  J. Langford,et al.  Reducing T-step reinforcement learning to classifica-tion , 2003 .

[26]  Nicholas Roy,et al.  Rapidly-exploring Random Belief Trees for motion planning under uncertainty , 2011, 2011 IEEE International Conference on Robotics and Automation.

[27]  Bernhard Nebel,et al.  Guiding the Generation of Manipulation Plans by Qualitative Spatial Reasoning , 2011, Spatial Cogn. Comput..

[28]  Emilio Frazzoli,et al.  Anytime Motion Planning using the RRT* , 2011, 2011 IEEE International Conference on Robotics and Automation.

[29]  Juan Chen,et al.  A survey of qualitative spatial representations , 2013, Knowledge engineering review (Print).

[30]  Milan Simic,et al.  Sampling-Based Robot Motion Planning: A Review , 2014, IEEE Access.