Reactive Execution of Learned Tasks With Real-Time Collision Avoidance in a Dynamic Environment

This paper addresses the problem of learning from demonstration (LfD) and subsequent robot safety control in an unstructured dynamic environment different from the demonstrations. Generally, LfD has been successfully exploited for task programming, but the existing methods have not solved the problem of allowing the entire arm to avoid obstacles while satisfying the task motion constraints (e.g., the robotic arm approaching the target object while avoiding obstacles moving within the environment). To achieve this, we present an incremental LfD approach that combines a task-parameterized probabilistic model and the robot security domain to control a robot’s behavior during task execution. Specifically, we propose a safety-oriented and task-oriented control strategy for redundant manipulators that makes full use of the motion redundancy of the manipulator and the space with no task restraints to satisfy the task constraints for human-robot coexistence. We then demonstrate the effectiveness of the proposed approach through a series of pick-and-pour experiments performed by a manipulator with 7 degree of freedom in a dynamic environment, where the robot must both avoid obstacles and satisfactorily complete the learned task with constraints.

[1]  Gregory D. Hager,et al.  An incremental approach to learning generalizable robot tasks from human demonstration , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Stefan Schaal,et al.  Learning feedback terms for reactive planning and control , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[4]  Jun Nakanishi,et al.  Dynamical Movement Primitives: Learning Attractor Models for Motor Behaviors , 2013, Neural Computation.

[5]  Emilio Frazzoli,et al.  Sampling-based algorithms for optimal motion planning , 2011, Int. J. Robotics Res..

[6]  Darwin G. Caldwell,et al.  A task-parameterized probabilistic model with minimal intervention control , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Mohamed Medhat Gaber,et al.  Imitation Learning , 2017, ACM Comput. Surv..

[8]  Alireza Akbarzadeh,et al.  Real-time velocity scaling and obstacle avoidance for industrial robots using fuzzy dynamic movement primitives and virtual impedances , 2018, Ind. Robot.

[9]  Marc Toussaint,et al.  Understanding the geometry of workspace obstacles in Motion Optimization , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[10]  Darwin G. Caldwell,et al.  Robot motor skill coordination with EM-based Reinforcement Learning , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[11]  Ron Alterovitz,et al.  Closed-loop global motion planning for reactive execution of learned tasks , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  E. AmirM.Ghalamzan,et al.  Robot learning from demonstrations: Emulation learning in environments with moving obstacles , 2018, Robotics Auton. Syst..

[13]  O. Khatib,et al.  Real-Time Obstacle Avoidance for Manipulators and Mobile Robots , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[14]  Bruno Siciliano,et al.  A closed-loop inverse kinematic scheme for on-line joint-based robot control , 1990, Robotica.

[15]  Ron Alterovitz,et al.  Asymptotically Optimal Motion Planning for Tasks Using Learned Virtual Landmarks , 2016, IEEE Robotics and Automation Letters.

[16]  Yuan Xing,et al.  Evaluation of robotic surgery skills using dynamic time warping , 2017, Comput. Methods Programs Biomed..

[17]  Jan Peters,et al.  Probabilistic Prioritization of Movement Primitives , 2017, IEEE Robotics and Automation Letters.

[18]  Ron Alterovitz,et al.  Asymptotically Optimal Motion Planning for Learned Tasks Using Time-Dependent Cost Maps , 2015, IEEE Transactions on Automation Science and Engineering.

[19]  Ajay Kumar Tanwani,et al.  Learning Robot Manipulation Tasks With Task-Parameterized Semitied Hidden Semi-Markov Model , 2016, IEEE Robotics and Automation Letters.

[20]  Siddhartha S. Srinivasa,et al.  CHOMP: Covariant Hamiltonian optimization for motion planning , 2013, Int. J. Robotics Res..

[21]  Jan Peters,et al.  Guiding Trajectory Optimization by Demonstrated Distributions , 2017, IEEE Robotics and Automation Letters.

[22]  Zengxi Pan,et al.  Recent progress on sampling based dynamic motion planning algorithms , 2016, 2016 IEEE International Conference on Advanced Intelligent Mechatronics (AIM).

[23]  Stefan Schaal,et al.  Biologically-inspired dynamical systems for movement generation: Automatic real-time goal adaptation and obstacle avoidance , 2009, 2009 IEEE International Conference on Robotics and Automation.

[24]  Nikolaos G. Tsagarakis,et al.  Statistical dynamical systems for skills acquisition in humanoids , 2012, 2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012).

[25]  Jan Peters,et al.  Using probabilistic movement primitives in robotics , 2017, Autonomous Robots.

[26]  Scott Niekum,et al.  Learning grounded finite-state representations from unstructured demonstrations , 2015, Int. J. Robotics Res..

[27]  Jan Peters,et al.  Phase estimation for fast action recognition and trajectory generation in human–robot collaboration , 2017, Int. J. Robotics Res..

[28]  Andrea Maria Zanchettin,et al.  Safety Assessment and Control of Robotic Manipulators Using Danger Field , 2013, IEEE Transactions on Robotics.

[29]  Anders Robertsson,et al.  Autonomous interpretation of demonstrations for modification of dynamical movement primitives , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).