Robot learning such as reinforcement learning generally needs a well-defined state space in order to converge. However, to build such a state space is one of the main issues of the robot learning because of the inter-dependence between state and action spaces, which resembles to the well known“chicken and egg”problem. This paper proposes a method of action-based state space construction for vision-based mobile robots. Basic ideas to cope with the interdependence are that we define a state as a cluster of input vectors from which the robot can reach the goal state or the state already obtained by a sequence of one kind action primitive regardless of its length, and that this sequence is defined as one action. To realize these ideas, we need many data (experiences) of the robot and cluster the input vectors as hyper ellipsoids so that the whole state space is segmented into a state transition map in terms of action from which the optimal action sequence is obtained. To show the validity of the method, we apply it to a soccer robot which tries to shoot a ball into a goal. The simulation and real experiments are shown.
[1]
Leslie Pack Kaelbling,et al.
Input Generalization in Delayed Reinforcement Learning: An Algorithm and Performance Comparisons
,
1991,
IJCAI.
[2]
Patrick Reignier,et al.
Learning to categorize perceptual space of a mobile robot using fuzzy-ART neural network
,
1994,
Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS'94).
[3]
Minoru Asada,et al.
Motion Sketch: Acquisition of Visual Motion Guided Behaviors
,
1995,
IJCAI.
[4]
Minoru Asada,et al.
Stereo sketch: stereo vision-based target reaching behavior acquisition with occlusion detection and avoidance
,
1996,
Proceedings of IEEE International Conference on Robotics and Automation.