Learning to dress

Creating animation of a character putting on clothing is challenging due to the complex interactions between the character and the simulated garment. We take a model-free deep reinforcement learning (deepRL) approach to automatically discovering robust dressing control policies represented by neural networks. While deepRL has demonstrated several successes in learning complex motor skills, the data-demanding nature of the learning algorithms is at odds with the computationally costly cloth simulation required by the dressing task. This paper is the first to demonstrate that, with an appropriately designed input state space and a reward function, it is possible to incorporate cloth simulation in the deepRL framework to learn a robust dressing control policy. We introduce a salient representation of haptic information to guide the dressing process and utilize it in the reward function to provide learning signals during training. In order to learn a prolonged sequence of motion involving a diverse set of manipulation skills, such as grasping the edge of the shirt or pulling on a sleeve, we find it necessary to separate the dressing task into several subtasks and learn a control policy for each subtask. We introduce a policy sequencing algorithm that matches the distribution of output states from one task to the input distribution for the next task in the sequence. We have used this approach to produce character controllers for several dressing tasks: putting on a t-shirt, putting on a jacket, and robot-assisted dressing of a sleeve.

[1]  Libin Liu,et al.  Guided Learning of Control Graphs for Physics-Based Characters , 2016, ACM Trans. Graph..

[2]  Yuyu Xu,et al.  Towards Cloth-Manipulating Characters , 2014, CASA 2014.

[3]  Takamitsu Matsubara,et al.  Reinforcement learning of clothing assistance with a dual-arm robot , 2011, 2011 11th IEEE-RAS International Conference on Humanoid Robots.

[4]  C. Karen Liu,et al.  Learning symmetric and low-energy locomotion , 2018, ACM Trans. Graph..

[5]  C. Karen Liu,et al.  Dexterous Manipulation of Cloth , 2016, Comput. Graph. Forum.

[6]  Libin Liu,et al.  Learning to schedule control fragments for physics-based characters using deep Q-learning , 2017, TOGS.

[7]  C. Karen Liu,et al.  Data-Driven Approach to Simulating Realistic Human Joint Constraints , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[8]  Taku Komura,et al.  Character Motion Synthesis by Topology Coordinates , 2009, Comput. Graph. Forum.

[9]  Tae-Yong Kim,et al.  Unified particle physics for real-time applications , 2014, ACM Trans. Graph..

[10]  Sergey Levine,et al.  Trust Region Policy Optimization , 2015, ICML.

[11]  Kourosh Naderi,et al.  Discovering and synthesizing humanoid climbing movements , 2017, ACM Trans. Graph..

[12]  Dmitry Berenson,et al.  Manipulation of deformable objects without modeling and simulating deformation , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Trevor Darrell,et al.  A geometric approach to robotic laundry folding , 2012, Int. J. Robotics Res..

[14]  Demis Hassabis,et al.  Mastering the game of Go without human knowledge , 2017, Nature.

[15]  Martin de Lasa,et al.  Feature-based locomotion controllers , 2010, ACM Trans. Graph..

[16]  Sergey Levine,et al.  High-Dimensional Continuous Control Using Generalized Advantage Estimation , 2015, ICLR.

[17]  Michiel van de Panne,et al.  Flexible muscle-based locomotion for bipedal creatures , 2013, ACM Trans. Graph..

[18]  Siddhartha S. Srinivasa,et al.  DART: Dynamic Animation and Robotics Toolkit , 2018, J. Open Source Softw..

[19]  Pieter Abbeel,et al.  Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.

[20]  Yuval Tassa,et al.  Emergence of Locomotion Behaviours in Rich Environments , 2017, ArXiv.

[21]  C. Karen Liu,et al.  Animating human dressing , 2015, ACM Trans. Graph..

[22]  David C. Brogan,et al.  Animating human athletics , 1995, SIGGRAPH.

[23]  Pieter Abbeel,et al.  Gravity-Based Robotic Cloth Folding , 2010, WAFR.

[24]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[25]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[26]  Stefano Carpin,et al.  Motion planning for cooperative manipulators folding flexible planar objects , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[27]  Yuval Tassa,et al.  Simulation tools for model-based robotics: Comparison of Bullet, Havok, MuJoCo, ODE and PhysX , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[28]  C. Karen Liu,et al.  Optimization-based interactive motion synthesis , 2009, ACM Trans. Graph..

[29]  Greg Turk,et al.  Learning to navigate cloth using haptics , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[30]  Sergey Levine,et al.  DeepMimic , 2018, ACM Trans. Graph..

[31]  Taku Komura,et al.  Harmonic parameterization by electrostatics , 2013, TOGS.

[32]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[33]  KangKang Yin,et al.  SIMBICON: simple biped locomotion control , 2007, ACM Trans. Graph..