Repeatable Folding Task by Humanoid Robot Worker Using Deep Learning

We propose a practical state-of-the-art method to develop a machine-learning-based humanoid robot that can work as a production line worker. The proposed approach provides an intuitive way to collect data and exhibits the following characteristics: task performing capability, task reiteration ability, generalizability, and easy applicability. The proposed approach utilizes a real-time user interface with a monitor and provides a first-person perspective using a head-mounted display. Through this interface, teleoperation is used for collecting task operating data, especially for tasks that are difficult to be applied with a conventional method. A two-phase deep learning model is also utilized in the proposed approach. A deep convolutional autoencoder extracts images features and reconstructs images, and a fully connected deep time delay neural network learns the dynamics of a robot task process from the extracted image features and motion angle signals. The “Nextage Open” humanoid robot is used as an experimental platform to evaluate the proposed model. The object folding task utilizing with 35 trained and 5 untrained sensory motor sequences for test. Testing the trained model with online generation demonstrates a 77.8% success rate for the object folding task.

[1]  Geoffrey E. Hinton,et al.  A time-delay neural network architecture for isolated word recognition , 1990, Neural Networks.

[2]  Tetsuya Ogata,et al.  Human-robot cooperation in arrangement of objects using confidence measure of neuro-dynamical system , 2010, 2010 IEEE International Conference on Systems, Man and Cybernetics.

[3]  Kyoichi Tatsuno,et al.  Task performance tests on inserting the bolts by an experimental system for power distribution line maintenance - grope action under compliance control , 2012, 2012 International Symposium on Micro-NanoMechatronics and Human Science (MHS).

[4]  Trevor Darrell,et al.  A geometric approach to robotic laundry folding , 2012, Int. J. Robotics Res..

[5]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[6]  J. Šindler World representation for a dual-arm robot manipulating with clothes , 2013 .

[7]  Shimon Y. Nof,et al.  Handbook of Industrial Robotics , 2013 .

[8]  Yuki Suga,et al.  Multimodal integration learning of robot behavior using deep neural networks , 2014, Robotics Auton. Syst..

[9]  Honglak Lee,et al.  Deep learning for detecting robotic grasps , 2013, Int. J. Robotics Res..

[10]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[11]  Mamoru Minami,et al.  Shape modeling of a string and recognition using distance sensor , 2015, 2015 24th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[12]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[13]  Kimitoshi Yamazaki,et al.  Wiring with Hooking of a String by a Dual-Armed Robot , 2016 .

[14]  Xiaoou Tang,et al.  Image Super-Resolution Using Deep Convolutional Networks , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Sergey Levine,et al.  End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..

[16]  Gordon Cheng,et al.  Motion Generation of Flexible Object Folding Task applied on Humanid Robot Using Deep Learning , 2016 .

[17]  Abhinav Gupta,et al.  Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Sergey Levine,et al.  Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection , 2016, Int. J. Robotics Res..

[19]  Chainer , 2019, Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining.