Reinforcement learning of a motor skill for wearing a T-shirt using topology coordinates

This article focuses on learning motor skills for anthropomorphic robots that must interact with non-rigid materials to perform such tasks as wearing clothes, turning socks inside out, and applying bandages. We propose a novel reinforcement learning framework for learning motor skills that interact with non-rigid materials. Our learning framework focuses on the topological relationship between the configuration of the robot and the non-rigid material based on the consideration that most details of the material (e.g. wrinkles) are not important for performing the motor tasks. This focus allows us to define the task performance and provide reward signals based on a low-dimensional variable, i.e. topology coordinates, in a real environment using reliable sensors. We constructed an experimental setting with an anthropomorphic dual-arm robot and a tailor-made T-shirt for it. To demonstrate the feasibility of our framework, a robot performed a T-shirt wearing task, whose goal was to put both of its arms into the corresponding sleeves of the T-shirt. The robot acquired sequential movements that put both of its arms into the T-shirt.

[1]  Jun Morimoto,et al.  Learning CPG-based Biped Locomotion with a Policy Gradient Method: Application to a Humanoid Robot , 2005, 5th IEEE-RAS International Conference on Humanoid Robots, 2005..

[2]  Kazuhito Yokoi,et al.  Leg motion primitives for a dancing humanoid robot , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[3]  Gordon Cheng,et al.  Learning tasks from observation and practice , 2004, Robotics Auton. Syst..

[4]  Nobuyuki Kita,et al.  Clothes state recognition using 3D observed data , 2009, 2009 IEEE International Conference on Robotics and Automation.

[5]  John N. Tsitsiklis,et al.  Actor-Critic Algorithms , 1999, NIPS.

[6]  Stefano Carpin,et al.  Combining imitation and reinforcement learning to fold deformable planar objects , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Jun Morimoto,et al.  Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning , 2000, Robotics Auton. Syst..

[8]  Jun Morimoto,et al.  Learning to acquire whole-body humanoid CoM movements to achieve dynamic tasks , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[9]  Jun Morimoto,et al.  Learning to Acquire Whole-Body Humanoid Center of Mass Movements to Achieve Dynamic Tasks , 2008, Adv. Robotics.

[10]  Stefan Schaal,et al.  2008 Special Issue: Reinforcement learning of motor skills with policy gradients , 2008 .

[11]  Mitsuo Kawato,et al.  A theory for cursive handwriting based on the minimization principle , 1995, Biological Cybernetics.

[12]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[13]  Jun Morimoto,et al.  Learning CPG-based biped locomotion with a policy gradient method , 2005, 5th IEEE-RAS International Conference on Humanoid Robots, 2005..

[14]  Yasuharu Koike,et al.  PII: S0893-6080(96)00043-3 , 1997 .

[15]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[16]  Kimitoshi Yamazaki,et al.  A Cloth Detection Method Based on Image Wrinkle Feature for Daily Assistive Robots , 2009, MVA.

[17]  Tsuneo Yoshikawa,et al.  Acquisition of a page turning skill for a multifingered hand using reinforcement learning , 2004, Adv. Robotics.

[18]  H. Sebastian Seung,et al.  Stochastic policy gradient reinforcement learning on a simple 3D biped , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[19]  Pieter Abbeel,et al.  Cloth grasp point detection based on multiple-view geometric cues with application to robotic towel folding , 2010, 2010 IEEE International Conference on Robotics and Automation.

[20]  Taku Komura,et al.  Character Motion Synthesis by Topology Coordinates , 2009, Comput. Graph. Forum.