Learning from demonstration with partially observable task parameters using dynamic movement primitives and Gaussian process regression

The problem of learning from demonstration in the case of partially observable external task parameters is addressed in this paper. Such a situation could be present in the daily life scenarios, where information regarding some task parameters are missing or partially available. In the first step, one dynamic movement primitives (DMP) model is learned for each demonstration trajectory. The parameters of the learned DMP model are recorded together with the corresponding external task parameters (query points), to create a database. Then Gaussian process regression (GPR) is used to create a model for the external task parameters and the DMP parameters. During reproduction, DMP parameters are retrieved by providing the new external task parameters and are used to regenerate the trajectory. It is shown how the learning approach could be adapted to deal with the partially observable external task parameters and regenerate the proper trajectory. The proposed methodology is applied to learn a via-point passing experiment with a lightweight robot manipulator (KUKA robot) to illustrate the efficacy of the proposed approach.

[1]  Jun Morimoto,et al.  Learning parametric dynamic movement primitives from multiple demonstrations , 2011, Neural Networks.

[2]  Darwin G. Caldwell,et al.  Learning from demonstrations with partially observable task parameters , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[4]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[5]  Aaron Hertzmann,et al.  Style machines , 2000, SIGGRAPH 2000.

[6]  Jan Peters,et al.  Nonamemanuscript No. (will be inserted by the editor) Reinforcement Learning to Adjust Parametrized Motor Primitives to , 2011 .

[7]  Henk Nijmeijer,et al.  Robot Programming by Demonstration , 2010, SIMPAR.

[8]  Stefan Schaal,et al.  Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.

[9]  Jun Nakanishi,et al.  Dynamical Movement Primitives: Learning Attractor Models for Motor Behaviors , 2013, Neural Computation.

[10]  Danica Kragic,et al.  Learning Actions from Observations , 2010, IEEE Robotics & Automation Magazine.

[11]  Darwin G. Caldwell,et al.  On improving the extrapolation capability of task-parameterized movement models , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  Jun Morimoto,et al.  On-line motion synthesis and adaptation using a trajectory database , 2012, Robotics Auton. Syst..

[13]  Nikos G. Tsagarakis,et al.  Teaching of bimanual skills in a compliant humanoid robot , 2012 .

[14]  Jun Morimoto,et al.  Real-time stylistic prediction for whole-body human motions , 2012, Neural Networks.

[15]  Jun Morimoto,et al.  Task-Specific Generalization of Discrete and Periodic Dynamic Movement Primitives , 2010, IEEE Transactions on Robotics.

[16]  Aude Billard,et al.  On Learning, Representing, and Generalizing a Task in a Humanoid Robot , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[17]  Olivier Sigaud,et al.  Learning compact parameterized skills with a single regression , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).

[18]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[19]  Michael I. Jordan,et al.  Supervised learning from incomplete data via an EM approach , 1993, NIPS.

[20]  Nikolaos G. Tsagarakis,et al.  Statistical dynamical systems for skills acquisition in humanoids , 2012, 2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012).

[21]  Klas Kronander,et al.  Learning to control planar hitting motions in a minigolf-like task , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.