Autonomous agent based on reinforcement learning and adaptive shadowed network

Abstract The planning of intelligent robot behavior plays an important role in the development of flexible automated systems. The robot’s intelligence comprises its capability to act in unpredictable and chaotic situations, which requires not just a change but the creation of the robot’s working knowledge. Planning of intelligent robot behavior addresses three main issues: finding task solutions in unknown situations, learning from experience and recognizing the similarity of problem paradigms. This article outlines a planning system which integrates the reinforcement learning method and a neural network approach with the aim to ensure autonomous robot behavior in unpredictable working conditions. The assumption is that the robot is a tabula rasa and has no knowledge of the work space structure. Initially, it has just basic strategic knowledge of searching for solutions, based on random attempts, and a built-in learning system. The reinforcement learning method is used here to evaluate robot behavior and to induce new, or improve the existing, knowledge. The acquired action (task) plan is stored as experience which can be used in solving similar future problems. To provide the recognition of problem similarities, the Adaptive Fuzzy Shadowed neural network is designed. This novel network concept with a fuzzy learning rule and shadowed hidden layer architecture enables the recognition of slightly translated or rotated patterns and does not forget already learned structures. The intelligent planning system is simulated using object-oriented techniques and verified on planned and random examples, proving the main advantages of the proposed approach: autonomous learning, which is invariant with regard to the order of training samples, and single iteration learning progress.

[1]  Sridhar Mahadevan,et al.  Automatic Programming of Behavior-Based Robots Using Reinforcement Learning , 1991, Artif. Intell..

[2]  Yoram Reich The development of Bridger: A methodological study of research on the use of machine learning in design , 1993, Artif. Intell. Eng..

[3]  Chuen-Sheng Cheng,et al.  A neural network-based cell formation algorithm in cellular manufacturing , 1995 .

[4]  Benjamin Kuipers,et al.  Learning hill-climbing functions as a strategy for generating behaviors in a mobile robot , 1991 .

[5]  Stephen Cameron,et al.  The Virtual Springs Method: Path Planning and Collision Avoidance for Redundant Manipulators , 1996, Int. J. Robotics Res..

[6]  Michael I. Jordan,et al.  Learning Fine Motion by Markov Mixtures of Experts , 1995, NIPS.

[7]  Nallan C. Suresh,et al.  Performance of Fuzzy ART neural network for group technology cell formation , 1994 .

[8]  Peihua Gu,et al.  A multi-constraint neural network for the pragmatic design of cellular manufacturing systems , 1995 .

[9]  James A. Freeman,et al.  Simulating neural networks - with Mathematica , 1993 .

[10]  Anil K. Jain,et al.  Artificial Neural Networks: A Tutorial , 1996, Computer.

[11]  Ben J. A. Kröse,et al.  A self-organizing representation of sensor space for mobile robot navigation , 1994, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS'94).

[12]  Sven Koenig,et al.  Complexity Analysis of Real-Time Reinforcement Learning , 1992, AAAI.

[13]  Timothy Masters,et al.  Practical neural network recipes in C , 1993 .