Learning cost function and trajectory for robotic writing motion

We present algorithms for inferring the cost function and reference trajectory from human demonstrations of hand-writing tasks. These two key elements are then used, through optimal control, to generate an impedance-based controller for a robotic hand. The key novelty lies in the flexibility of the feature design in the composition of the cost function, in contrast to the traditional approaches that consider linearly combined features. Cross-entropy-based methods form the core of our learning technique, resulting in sample-based stochastic algorithms for task encoding and decoding. The algorithms are validated using an anthropomorphic robot hand. We assess that the correct compliance is well encapsulated by subjecting the robot to perturbations during task reproduction.

[1]  Aude Billard,et al.  Learning to Play Minigolf: A Dynamical System-Based Approach , 2012, Adv. Robotics.

[2]  Darwin G. Caldwell,et al.  Multi-optima exploration with adaptive Gaussian mixture model , 2012, 2012 IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL).

[3]  Pierre-Yves Oudeyer,et al.  Adaptive exploration through covariance matrix adaptation enables developmental motor learning , 2012, Paladyn J. Behav. Robotics.

[4]  Marc Toussaint,et al.  Robot trajectory optimization using approximate inference , 2009, ICML '09.

[5]  Vicenç Gómez,et al.  Optimal control as a graphical model inference problem , 2009, Machine Learning.

[6]  Jun Nakanishi,et al.  Movement imitation with nonlinear dynamical systems in humanoid robots , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[7]  J. Andrew Bagnell,et al.  Maximum margin planning , 2006, ICML.

[8]  Shie Mannor,et al.  A Tutorial on the Cross-Entropy Method , 2005, Ann. Oper. Res..

[9]  A. Billard,et al.  Learning Stable Nonlinear Dynamical Systems With Gaussian Mixture Models , 2011, IEEE Transactions on Robotics.

[10]  Aude Billard,et al.  Learning object-level impedance control for robust grasping and dexterous manipulation , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[11]  Anind K. Dey,et al.  Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.

[12]  Marc Toussaint,et al.  Learned graphical models for probabilistic planning provide a new class of movement primitives , 2013, Front. Comput. Neurosci..

[13]  Aude Billard,et al.  Learning Stable Nonlinear Dynamical Systems With Gaussian Mixture Models , 2011, IEEE Transactions on Robotics.

[14]  Marin Kobilarov,et al.  Cross-entropy motion planning , 2012, Int. J. Robotics Res..

[15]  Nikolaos G. Tsagarakis,et al.  Statistical dynamical systems for skills acquisition in humanoids , 2012, 2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012).

[16]  Sergey Levine,et al.  Continuous Inverse Optimal Control with Locally Optimal Examples , 2012, ICML.