Exploration and model building in mobile robot domains

The first results on COLUMBUS, an autonomous mobile robot, are presented. COLUMBUS operates in initially unknown structured environments. Its task is to explore and model the environment efficiently while avoiding collisions with obstacles. COLUMBUS uses an instance-based learning technique for modeling its environment. Real-world experiences are generalized via two artificial neural networks that encode the characteristics of the robot's sensors, as well as the characteristics of typical environments which the robot is assumed to face. Once trained, these networks allow for the transfer of knowledge across different environments the robot will face over its lifetime. Exploration is achieved by navigating to low confidence regions. A dynamic programming method is employed in background to find minimal-cost paths that, when executed by the robot, maximize exploration.<<ETX>>

[1]  G. D. Liveing,et al.  The University of Cambridge , 1897, British medical journal.

[2]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[3]  Alberto Elfes,et al.  Sonar-based real-world mapping and navigation , 1987, IEEE J. Robotics Autom..

[4]  Hans P. Moravec Sensor Fusion in Certainty Grids for Mobile Robots , 1988, AI Mag..

[5]  Andrew W. Moore,et al.  Efficient memory-based learning for robot control , 1990 .

[6]  Richard S. Sutton,et al.  Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.

[7]  Jonathan Bachrach,et al.  A Connectionist Learning Control Architecture for Navigation , 1990, NIPS.

[8]  Michael I. Jordan,et al.  A Modular Connectionist Architecture For Learning Piecewise Control Strategies , 1991, 1991 American Control Conference.

[9]  Dieter Fox,et al.  Learning By Error-Driven Decomposition , 1991 .

[10]  J. Friedman Multivariate adaptive regression splines , 1990 .

[11]  Christopher G. Atkeson,et al.  Using locally weighted regression for robot learning , 1991, Proceedings. 1991 IEEE International Conference on Robotics and Automation.

[12]  Long Ji Lin,et al.  Programming Robots Using Reinforcement Learning and Teaching , 1991, AAAI.

[13]  Sebastian Thrun,et al.  Active Exploration in Dynamic Environments , 1991, NIPS.

[14]  J. Freidman,et al.  Multivariate adaptive regression splines , 1991 .

[15]  Satinder P. Singh,et al.  The Efficient Learning of Multiple Task Sequences , 1991, NIPS.

[16]  R.G. Simmons,et al.  Concurrent planning and execution for autonomous robots , 1992, IEEE Control Systems.

[17]  Sebastian Thrun,et al.  The role of exploration in learning control , 1992 .

[18]  Sebastian Thrun,et al.  Efficient Exploration In Reinforcement Learning , 1992 .

[19]  Donald A. Sofge,et al.  Handbook of Intelligent Control: Neural, Fuzzy, and Adaptive Approaches , 1992 .