Research And Development On Searching A Routing Path Of A Dynamic Terrain

The research topic of this paper is that how an avatar explores the environment, and find the way out to reach the pre-assigned goal. This proposed system is applied to a dynamicallly changable, continuous and largescaled environment. To deal with the learning in such environment, the proposed system partitions the statespace into regions of states, called cells. By using cells, the states of the system can be reduced in which the number of states grows dramatically increasing in proportion to the number and quality of inputs. Through a series of maniputations of cells, the avatar can adapt itself to the changable environment. By adjusting the peferences of the proposed global-cell, the tendancy of exploration behavior of the avatar can be controlled to explore the potential path caused by the changing environment.

[1]  Patrick K. Simpson,et al.  Fuzzy min-max neural networks - Part 2: Clustering , 1993, IEEE Trans. Fuzzy Syst..

[2]  David H. Wolpert,et al.  Bandit problems and the exploration/exploitation tradeoff , 1998, IEEE Trans. Evol. Comput..

[3]  Andrew W. Moore,et al.  The parti-game algorithm for variable resolution reinforcement learning in multidimensional state-spaces , 2004, Machine Learning.

[4]  Sean R Eddy,et al.  What is dynamic programming? , 2004, Nature Biotechnology.

[5]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[6]  P. K. Simpson,et al.  Fuzzy min-max neural networks , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.

[7]  Patrick K. Simpson,et al.  Fuzzy min-max neural networks. I. Classification , 1992, IEEE Trans. Neural Networks.