Reactive phase and task space adaptation for robust motion execution

An essential aspect for making robots succeed in real-world environments is to give them the ability to robustly perform motions in continuously changing situations. Classical motion planning methods usually create plans for static environments. The direct execution of such plans in dynamic environments often becomes problematic. We present an approach that adapts motion plans by feeding changes of the environment into a transformation of the plan in task space. Furthermore, the progress in the plan is defined with a phase variable that is updated adaptively according to the actual task progress. This phase variable releases the strict time compliance that many motion planning methods bring along. The main benefit of our approach is the ability to do this adaptation in a computational efficient manner during the execution of the motion. Thus, the gap between the motion planning and motion execution stage is bridged by continuously transforming geometric and dynamic features of a reference plan to the current situation. We evaluate the performance of our approach by comparing it to alternative methods such as dynamic motion primitives and continuous replanning on several simulated benchmark tasks. Moreover, we demonstrate the real robot applicability on a PR2 robot platform.

[1]  Eugene Fiume,et al.  Feedback control for rotational movements in feature space , 2014, Comput. Graph. Forum.

[2]  Igor Skrjanc,et al.  Tracking-error model-based predictive control for mobile robots in real time , 2007, Robotics Auton. Syst..

[3]  M. Spong,et al.  Robot Modeling and Control , 2005 .

[4]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[5]  Kazuhito Yokoi,et al.  Online replanning for reactive robot motion: Practical aspects , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[6]  Jan Peters,et al.  Imitation and Reinforcement Learning: Practical Algorithms for Motor Primitives in Robotics , 2010 .

[7]  J. Peters,et al.  Imitation and Reinforcement Learning â Practical Algorithms for Motor Primitive Learning in Robotics , 2010 .

[8]  Geoffrey Biggs,et al.  A Survey of Robot Programming Systems , 2010 .

[9]  Stefan Schaal,et al.  Dynamics systems vs. optimal control--a unifying view. , 2007, Progress in brain research.

[10]  Yi Gu,et al.  Space-indexed dynamic programming: learning to follow trajectories , 2008, ICML '08.

[11]  Christine Chevallereau Time-scaling control for an underactuated biped robot , 2003, IEEE Trans. Robotics Autom..

[12]  Jan Peters,et al.  Imitation and Reinforcement Learning , 2010, IEEE Robotics & Automation Magazine.

[13]  S. Schaal Dynamic Movement Primitives -A Framework for Motor Control in Humans and Humanoid Robotics , 2006 .

[14]  Carl E. Rasmussen,et al.  PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.

[15]  Dongjun Lee,et al.  Time-Scaling Trajectories of Passive-Dynamic Bipedal Robots , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[16]  Luis Moreno,et al.  Kinesthetic teaching via Fast Marching Square , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  O. Brock,et al.  Elastic Strips: A Framework for Motion Generation in Human Environments , 2002, Int. J. Robotics Res..

[18]  Nidhi Kalra,et al.  Replanning with RRTs , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[19]  Peter Englert,et al.  Probabilistic model-based imitation learning , 2013, Adapt. Behav..

[20]  Manfred Morari,et al.  Model predictive control: Theory and practice - A survey , 1989, Autom..

[21]  Oussama Khatib,et al.  A unified approach for motion and force control of robot manipulators: The operational space formulation , 1987, IEEE J. Robotics Autom..

[22]  Manfred Morari,et al.  Model predictive control: Theory and practice , 1988 .

[23]  Bruce A. MacDonald,et al.  Distance indexed trajectory generation for a helicopter robot for programming by demonstration , 2009, 2009 IEEE/ASME International Conference on Advanced Intelligent Mechatronics.

[24]  Darwin G. Caldwell,et al.  Handling of multiple constraints and motion alternatives in a robot programming by demonstration framework , 2009, 2009 9th IEEE-RAS International Conference on Humanoid Robots.

[25]  Pierre-Brice Wieber,et al.  Online walking gait generation with adaptive foot positioning through Linear Model Predictive control , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[26]  D Fox,et al.  Multiple-Target Reinforcement Learning with a Single Policy , 2011 .

[27]  Aude Billard,et al.  Dynamical System Modulation for Robot Learning via Kinesthetic Demonstrations , 2008, IEEE Transactions on Robotics.

[28]  Jan Peters,et al.  Probabilistic Movement Primitives , 2013, NIPS.

[29]  Oskar von Stryk,et al.  Direct and indirect methods for trajectory optimization , 1992, Ann. Oper. Res..

[30]  Stefan Schaal,et al.  2008 Special Issue: Reinforcement learning of motor skills with policy gradients , 2008 .