A study on vision-based mobile robot learning by deep Q-network

Deep Q-network (DQN) is one of the most famous methods of deep reinforcement learning. DQN approximates the action value function by Convolutional Neural Network (CNN) and updates the action value function by Q-learning. In this research, we apply DQN to robot behavior learning the simulation environment. We realize that the mobile robot learns to acquire good behaviors such as avoiding the wall and moving along the center line by using high-dimensional visual information as input data. We propose a method which reuses the best target network so far in case learning performance suddenly falls. Moreover, we incorporate Profit Sharing method to DQN in order to accelerate the learning. Through the simulation experiment, we confirm the effectiveness of our method.