A Practice Strategy for Robot Learning Control

"Trajectory Extension Learning" is a new technique for Learning Control in Robots which assumes that there exists some parameter of the desired trajectory that can be smoothly varied from a region of easy solvability of the dynamics to a region of desired behavior which may have more difficult dynamics. By gradually varying the parameter, practice movements remain near the desired path while a Neural Network learns to approximate the inverse dynamics. For example, the average speed of motion might be varied, and the inverse dynamics can be "bootstrapped" from slow movements with simpler dynamics to fast movements. This provides an example of the more general concept of a "Practice Strategy" in which a sequence of intermediate tasks is used to simplify learning a complex task. I show an example of the application of this idea to a real 2-joint direct drive robot arm.

[1]  Robert M. Sanner,et al.  Gaussian Networks for Direct Adaptive Control , 1991, 1991 American Control Conference.

[2]  Bo Egardt Adaptive control stability, convergence, and robustness : Shankar Sastry and Marc Bodson , 1993, Autom..

[3]  C. Atkeson,et al.  Learning arm kinematics and dynamics. , 1989, Annual review of neuroscience.

[4]  Suguru Arimoto,et al.  Bettering operation of Robots by learning , 1984, J. Field Robotics.

[5]  Takayuki Yamada,et al.  Learning control using neural networks , 1991, Proceedings. 1991 IEEE International Conference on Robotics and Automation.

[6]  S. Sastry,et al.  Adaptive Control: Stability, Convergence and Robustness , 1989 .

[7]  Terence D. Sanger,et al.  A Tree-Structured Algorithm for Reducing Computation in Networks with Separable Basis Functions , 1991, Neural Computation.

[8]  Reza Shadmehr Learning Virtual Equilibrium Trajectories for Control of a Robot Arm , 1990, Neural Computation.

[9]  E. Bizzi,et al.  Posture control and trajectory formation during arm movement , 1984, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[10]  Terence D. Sanger,et al.  A tree-structured adaptive network for function approximation in high-dimensional spaces , 1991, IEEE Trans. Neural Networks.

[11]  Terence D. Sanger,et al.  Neural network learning control of robot manipulators using gradually increasing task difficulty , 1994, IEEE Trans. Robotics Autom..