MGRL: Graph neural network based inference in a Markov network with reinforcement learning for visual navigation

Abstract Visual navigation is an essential task for indoor robots and usually uses the map as assistance to providing global information for the agent. Because the traditional maps match the environments, the map-based and map-building-based navigation methods are limited in the new environments for obtaining maps. Although the deep reinforcement learning navigation method, utilizing the non-map-based navigation technique, achieves satisfactory performance, it lacks the interpretability and the global view of the environment. Therefore, we propose a novel abstract map for the deep reinforcement learning navigation method with better global relative position information and more reasonable interpretability. The abstract map is modeled as a Markov network which is used for explicitly representing the regularity of objects arrangement, influenced by people activities in different environments. Besides, a knowledge graph is utilized to initialize the structure of the Markov network, as providing the prior structure for the model and reducing the difficulty of model learning. Then, a graph neural network is adopted for probability inference in the Markov network. Furthermore, the update of the abstract map, including the knowledge graph structure and the parameters of the graph neural network, are combined into an end-to-end learning process trained by a reinforcement learning method. Finally, experiments in the AI2THOR framework and the physical environment indicate that our algorithm greatly improves the success rate of navigation in case of new environments, thus confirming the good generalization.

[1]  H. Kappen Optimal control theory and the linear bellman equation , 2011 .

[2]  Sebastian Thrun,et al.  Learning Metric-Topological Maps for Indoor Mobile Robot Navigation , 1998, Artif. Intell..

[3]  Michael S. Bernstein,et al.  Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations , 2016, International Journal of Computer Vision.

[4]  Ah Chung Tsoi,et al.  The Graph Neural Network Model , 2009, IEEE Transactions on Neural Networks.

[5]  Enrique Onieva,et al.  Conditional Random Field-Based Offline Map Matching for Indoor Environments , 2016, Sensors.

[6]  Francisco Bonin-Font,et al.  Visual Navigation for Mobile Robots: A Survey , 2008, J. Intell. Robotic Syst..

[7]  Oussama Khatib,et al.  Real-Time Obstacle Avoidance for Manipulators and Mobile Robots , 1985, Autonomous Robot Vehicles.

[8]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[9]  Dongbin Zhao,et al.  Deep Reinforcement Learning With Visual Attention for Vehicle Classification , 2017, IEEE Transactions on Cognitive and Developmental Systems.

[10]  Ye Zhou,et al.  Hybrid Hierarchical Reinforcement Learning for online guidance and navigation with partial observability , 2019, Neurocomputing.

[11]  Heesung Jun,et al.  Vision-based location positioning using augmented reality for indoor navigation , 2008, IEEE Transactions on Consumer Electronics.

[12]  David Filliat,et al.  Map-based navigation in mobile robots: II. A review of map-learning and path-planning strategies , 2003, Cognitive Systems Research.

[13]  Hongyu Guo,et al.  Indoor Pedestrian Navigation Based on Conditional Random Field Algorithm , 2017, Micromachines.

[14]  Roland Siegwart,et al.  Feature extraction and scene interpretation for map-based navigation and map building , 1998, Other Conferences.

[15]  Dongbin Zhao,et al.  StarCraft Micromanagement With Reinforcement Learning and Curriculum Transfer Learning , 2018, IEEE Transactions on Emerging Topics in Computational Intelligence.

[16]  Hong Zhang,et al.  Path planning for intelligent robot based on switching local evolutionary PSO algorithm , 2016 .

[17]  Yoram Koren,et al.  Real-time obstacle avoidance for fact mobile robots , 1989, IEEE Trans. Syst. Man Cybern..