A New Approach in Agent Path-Finding using State Mark Gradients

Since searching is one of the most important problem-solving methods, especially in Artificial Intelligence where it is often difficult to devise straightforward solutions, it has been given continuous attention by researchers. In this paper a new algorithm for agent path-finding is presented. Our approach is based on environment marking during exploration. We tested the performances of Q-learning and Learning Real-Time A* algorithm for three proposed mazes and then a comparison was made between our algorithm, two variants of Q-learning and LRTA* algorithm.

[1]  Gerhard Weiss Search Algorithms for Agents , 2000 .

[2]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[3]  Bir Bhanu,et al.  Real-time robot learning , 2001, Proceedings 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164).

[4]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[5]  Michael Wooldridge,et al.  Multiagent Systems: A Modern Approach to Distributed Artificial Intelligence , 1999 .

[6]  Nils J. Nilsson,et al.  A Formal Basis for the Heuristic Determination of Minimum Cost Paths , 1968, IEEE Trans. Syst. Sci. Cybern..

[7]  Faruk Polat,et al.  Coordination of intelligent agents in real‐time search , 2002, Expert Syst. J. Knowl. Eng..

[8]  Richard E. Korf,et al.  Real-Time Heuristic Search , 1990, Artif. Intell..