Detecting insertion tasks using convolutional neural networks during robot teaching-by-demonstration

Today, collaborative robots are often taught new tasks through “teaching by demonstration” techniques rather than manual programming. This works well for many tasks; however, some tasks like precise tight-fitting insertions can be hard to recreate through exact position replays because they also involve forces and are highly affected by the robot's repeatability and the position of the object in the hand. As of yet there is no way to automatically detect when procedures to reduce position uncertainty should be used. In this paper, we present a new way to automatically detect insertion tasks during impedance control-based trajectory teaching. This is accomplished by recording the forces and torques applied by the operator and inputting these signals to a convolutional neural network. The convolutional neural network is used to extract important features of the spatio-temporal forces and torque signals for distinguishing insertion tasks. Eventually, this method could help robots understand the tasks they are taught at a higher level. They will not only be capable of a position-time replay of the task, but will also recognize the best strategy to apply in order to accomplish the task (in this case insertion). Our method was tested on data obtained from 886 experiments that were conducted on eight different in-hand objects. Results show that we can distinguish insertion tasks from pick-and-place tasks with an average accuracy of 82%.

[1]  Jan F. Broenink,et al.  Peg-in-Hole assembly using Impedance Control with a 6 DOF Robot , 1996 .

[2]  Aude Billard,et al.  Dynamical System Modulation for Robot Learning via Kinesthetic Demonstrations , 2008, IEEE Transactions on Robotics.

[3]  Sven Behnke,et al.  Evaluation of Pooling Operations in Convolutional Architectures for Object Recognition , 2010, ICANN.

[4]  Gerald Penn,et al.  Convolutional Neural Networks for Speech Recognition , 2014, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[5]  Koen E. A. van de Sande,et al.  Selective Search for Object Recognition , 2013, International Journal of Computer Vision.

[6]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[7]  Clément Gosselin,et al.  Safe, Stable and Intuitive Control for Physical Human-Robot Interaction , 2009, 2009 IEEE International Conference on Robotics and Automation.

[8]  Reid G. Simmons,et al.  An autonomous mobile manipulator for assembly tasks , 2010, Auton. Robots.

[9]  Antoni B. Chan,et al.  Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network , 2010 .

[10]  Donald R. Myers,et al.  Automated programming of an industrial robot through teach-by showing , 2001, Proceedings 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164).

[11]  Sebastian Böck,et al.  Improved musical onset detection with Convolutional Neural Networks , 2014, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[12]  Stefan Schaal,et al.  Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.

[13]  Hiroshi Imamizu,et al.  Human cerebellar activity reflecting an acquired internal model of a new tool , 2000, Nature.

[14]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[15]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[16]  Geoffrey E. Hinton,et al.  Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[17]  Tony Wong,et al.  Unsupervised feature learning for classifying dynamic tactile events using sparse coding , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Honglak Lee,et al.  Unsupervised feature learning for audio classification using convolutional deep belief networks , 2009, NIPS.

[19]  Tadej Petric,et al.  Human-in-the-loop approach for teaching robot assembly tasks using impedance control interface , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[20]  Thomas Deselaers,et al.  Measuring the Objectness of Image Windows , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Kae Nakamura,et al.  Central mechanisms of motor skill learning , 2002, Current Opinion in Neurobiology.

[22]  Cesare Furlanello,et al.  Convolutional Neural Network for Stereotypical Motor Movement Detection in Autism , 2015, ArXiv.

[23]  Holger Voos,et al.  Position Identification in Force-Guided Robotic Peg-in-Hole Assembly Tasks , 2014 .

[24]  K. Doya Complementary roles of basal ganglia and cerebellum in learning and motor control , 2000, Current Opinion in Neurobiology.