Fast and Stable Learning in Direct-Vision-Based Reinforcement learning

Direct-Vision-Based Reinforcement Learning has been proposed not only for the motion planning but for the learning of the whole process from sensors to motors in robots, including recognition, attention and so on. In this learning, raw visual sensory signals are put into a layered neural network directly, and the network is trained by the training signals generated based on reinforcement learning. On the other hand, it has been pointed out that the combination of neural network and TD-type reinforcement learning sometimes leads to instability of learning. In this paper, it is shown that each visual sensory cell makes a role of localization of our continuous 3-dimensional space and it helps the learning to be fast and stable. Further by processing the localized input signals in the layered neural network, a global representation is reconstructed adaptively in the hidden layer through learning as shown in the previous papers.