A Hybrid Architecture for Learning Robot Control Tasks

Autonomous robot systems operating in an uncertain environment have to be reactive and adaptive in order to cope with changing environment conditions and task requirements. To achieve this, the control architecture presented in this paper uses reinforcement learning on top of an abstract Discrete Event Dynamic System (DEDS) supervisor to learn to coordinate a set of continuous controllers in order to perform a given task. In addition to providing a base reactivity through the underlying stable and convergent control elements, the use of this hybrid control approach also allows the learning to be performed on an abstract system model which dramatically reduces the complexity of the learning problem. Furthermore, the DEDS formalism provides means of imposing safety constraints a priori, such that learning can be performed on-line in a single trial without the need for an outside teacher. To demonstrate the applicability of this approach, the architecture is used to learn a turning gait on a fourlegged robot platform.

[1]  Richard S. Sutton,et al.  Neuronlike adaptive elements that can solve difficult learning control problems , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[2]  W. M. Wonham,et al.  The control of discrete event systems , 1989 .

[3]  Rodney A. Brooks,et al.  Learning to Coordinate Behaviors , 1990, AAAI.

[4]  吉川 恒夫,et al.  Foundations of robotics : analysis and control , 1990 .

[5]  Vijaykumar Gullapalli,et al.  Learning Control Under Extreme Uncertainty , 1992, NIPS.

[6]  Long-Ji Lin,et al.  Reinforcement learning for robots using neural networks , 1992 .

[7]  Sridhar Mahadevan,et al.  Automatic Programming of Behavior-Based Robots Using Reinforcement Learning , 1991, Artif. Intell..

[8]  Roderic A. Grupen,et al.  The applications of harmonic functions to robotics , 1993, J. Field Robotics.

[9]  Jana Kosecka,et al.  Application of discrete events systems for modeling and controlling robotic agents , 1994, Proceedings of the 1994 IEEE International Conference on Robotics and Automation.

[10]  Kimon P. Valavanis,et al.  A subject-indexed bibliography of discrete event dynamic systems , 1994, IEEE Robotics & Automation Magazine.

[11]  Andrew G. Barto,et al.  Improving Elevator Performance Using Reinforcement Learning , 1995, NIPS.

[12]  Roderic A. Grupen,et al.  Distributed Control Representation for Manipulation Tasks , 1995, IEEE Expert.

[13]  Andrew G. Barto,et al.  Learning to Act Using Real-Time Dynamic Programming , 1995, Artif. Intell..

[14]  José del R. Millán,et al.  Rapid, safe, and incremental learning of navigation strategies , 1996, IEEE Trans. Syst. Man Cybern. Part B.

[15]  Leslie Pack Kaelbling,et al.  On reinforcement learning for robots , 1996, IROS.

[16]  R. A. Grupen,et al.  TITLE A Hybrid Discrete Event Dynamic Systems Approach to Robot Control , 1996 .

[17]  Panos J. Antsaklis,et al.  A logical DES approach to the design of hybrid control systems , 1996 .

[18]  J. A. Coelho,et al.  A Control Basis for Learning Multifingered Grasps , 1997 .

[19]  J. A. Coelho,et al.  A control basis for learning multifingered grasps , 1997, J. Field Robotics.

[20]  Roderic A. Grupen,et al.  A Control Structure For Learning Locomotion Gaits , 1998 .

[21]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.