Robot Obstacle Avoidance Learning Based on Mixture Models

We briefly surveyed the existing obstacle avoidance algorithms; then a new obstacle avoidance learning framework based on learning from demonstration (LfD) is proposed. The main idea is to imitate the obstacle avoidance mechanism of human beings, in which humans learn to make a decision based on the sensor information obtained by interacting with environment. Firstly, we endow robots with obstacle avoidance experience by teaching them to avoid obstacles in different situations. In this process, a lot of data are collected as a training set; then, to encode the training set data, which is equivalent to extracting the constraints of the task, Gaussian mixture model (GMM) is used. Secondly, a smooth obstacle-free path is generated by Gaussian mixture regression (GMR). Thirdly, a metric of imitation performance is constructed to derive a proper control policy. The proposed framework shows excellent generalization performance, which means that the robots can fulfill obstacle avoidance task efficiently in a dynamic environment. More importantly, the framework allows learning a wide variety of skills, such as grasp and manipulation work, which makes it possible to build a robot with versatile functions. Finally, simulation experiments are conducted on a Turtlebot robot to verify the validity of our algorithms.

[1]  Jin-Hui Zhu,et al.  Obstacle avoidance with multi-objective optimization by PSO in dynamic environment , 2005, 2005 International Conference on Machine Learning and Cybernetics.

[2]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[3]  O. Khatib,et al.  Real-Time Obstacle Avoidance for Manipulators and Mobile Robots , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[4]  Tae-Yong Kuc,et al.  Genetic algorithm based path planning and dynamic obstacle avoidance of mobile robots , 1997, 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation.

[5]  Nancy Arana-Daniel,et al.  Obstacle avoidance using PSO , 2012, 2012 9th International Conference on Electrical Engineering, Computing Science and Automatic Control (CCE).

[6]  Yunong Zhang,et al.  Obstacle avoidance for kinematically redundant manipulators using a dual neural network , 2004, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[7]  Yoram Koren,et al.  Real-time obstacle avoidance for fast mobile robots in cluttered environments , 1990, Proceedings., IEEE International Conference on Robotics and Automation.

[8]  Min Guo,et al.  Reinforcement Learning Neural Network to the Problem of Autonomous Mobile Robot Obstacle Avoidance , 2005, 2005 International Conference on Machine Learning and Cybernetics.

[9]  Chrystopher L. Nehaniv,et al.  Like Me?- Measures of Correspondence and Imitation , 2001, Cybern. Syst..

[10]  N. Peric,et al.  A reinforcement learning approach to obstacle avoidance of mobile robots , 2002, 7th International Workshop on Advanced Motion Control. Proceedings (Cat. No.02TH8623).

[11]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[12]  Ashutosh Saxena,et al.  High speed obstacle avoidance using monocular vision and reinforcement learning , 2005, ICML.

[13]  Claude F. Touzet,et al.  Neural reinforcement learning for behaviour synthesis , 1997, Robotics Auton. Syst..

[14]  N. H. C. Yung,et al.  A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance , 2003, IEEE Trans. Syst. Man Cybern. Part B.

[15]  Jean-Claude Latombe,et al.  Robot motion planning , 1970, The Kluwer international series in engineering and computer science.

[16]  Ellips Masehian,et al.  Classic and Heuristic Approaches in Robot Motion Planning A Chronological Review , 2007 .

[17]  Yantao Tian,et al.  A Hybrid Ant Colony Optimization Algorithm for Path Planning of Robot in Dynamic Environment , 2006 .

[18]  Oscar Castillo,et al.  Path planning for autonomous mobile robot navigation with ant colony optimization and fuzzy cost function evaluation , 2009, Appl. Soft Comput..

[19]  P. Raja,et al.  Optimal path planning of mobile robots: A review , 2012 .

[20]  Yann LeCun,et al.  Off-Road Obstacle Avoidance through End-to-End Learning , 2005, NIPS.

[21]  Leslie Pack Kaelbling,et al.  Effective reinforcement learning for mobile robots , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[22]  Jelena Godjevac A learning procedure for a fuzzy system : application to obstacleavoidanceJelena , 1995 .

[23]  Richard S. Sutton,et al.  Introduction to Reinforcement Learning , 1998 .

[24]  Sourav Dutta Obstacle Avoidance of mobile robot using PSO based Neuro Fuzzy Technique , 2010 .

[25]  Simon X. Yang,et al.  Genetic algorithm based path planning for a mobile robot , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).