Random neural Q-learning for obstacle avoidance of a mobile robot in unknown environments

The article presents a random neural Q-learning strategy for the obstacle avoidance problem of an autonomous mobile robot in unknown environments. In the proposed strategy, two independent modules, namely, avoidance without considering the target and goal-seeking without considering obstacles, are first trained using the proposed random neural Q-learning algorithm to obtain their best control policies. Then, the two trained modules are combined based on a switching function to realize the obstacle avoidance in unknown environments. For the proposed random neural Q-learning algorithm, a single-hidden layer feedforward network is used to approximate the Q-function to estimate the Q-value. The parameters of the single-hidden layer feedforward network are modified using the recently proposed neural algorithm named the online sequential version of extreme learning machine, where the parameters of the hidden nodes are assigned randomly and the sample data can come one by one. However, different from the original online sequential version of extreme learning machine algorithm, the initial output weights are estimated subjected to quadratic inequality constraint to improve the convergence speed. Finally, the simulation results demonstrate that the proposed random neural Q-learning strategy can successfully solve the obstacle avoidance problem. Also, the higher learning efficiency and better generalization ability are achieved by the proposed random neural Q-learning algorithm compared with the Q-learning based on the back-propagation method.

[1]  Xiaogang Ruan,et al.  Q-learning Based on Neural Network in Learning Action Selection of Mobile Robot , 2007, 2007 IEEE International Conference on Automation and Logistics.

[2]  Kaustubh Pathak,et al.  An integrated path-planning and control approach for nonholonomic unicycles using switched local potentials , 2005, IEEE Transactions on Robotics.

[3]  Yoram Koren,et al.  Histogramic in-motion mapping for mobile robot obstacle avoidance , 1991, IEEE Trans. Robotics Autom..

[4]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[5]  Robert M. Sanner,et al.  Gaussian Networks for Direct Adaptive Control , 1991, 1991 American Control Conference.

[6]  Kaustubh Pathak,et al.  An Integrated Path Planning and Control Framework for Nonholonomic Unicycles , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[7]  Jingyuan Zhang,et al.  Application of Artificial Neural Network Based on Q-learning for Mobile Robot Path Planning , 2006, 2006 IEEE International Conference on Information Acquisition.

[8]  Velappa Ganapathy,et al.  Neural Q-Learning controller for mobile robot , 2009, 2009 IEEE/ASME International Conference on Advanced Intelligent Mechatronics.

[9]  Badong Chen,et al.  Quantized Kernel Least Mean Square Algorithm , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[10]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[11]  Nan Liu,et al.  Landmark recognition with sparse representation classification and extreme learning machine , 2015, J. Frankl. Inst..

[12]  Narasimhan Sundararajan,et al.  Classification of Mental Tasks from Eeg Signals Using Extreme Learning Machine , 2006, Int. J. Neural Syst..

[13]  Mark H. Overmars,et al.  Roadmap-based motion planning in dynamic environments , 2005, IEEE Trans. Robotics.

[14]  Chee Keong Kwoh,et al.  Extreme Learning Machine for Predicting HLA-Peptide Binding , 2006, ISNN.

[15]  Kimon P. Valavanis,et al.  Autonomous vehicle navigation utilizing electrostatic potential fields and fuzzy logic , 2001, IEEE Trans. Robotics Autom..

[16]  Xiaogang Ruan,et al.  Application of reinforcement learning based on neural network to dynamic obstacle avoidance , 2008, 2008 International Conference on Information and Automation.

[17]  Jiuwen Cao,et al.  Protein Sequence Classification with Improved Extreme Learning Machine Algorithms , 2014, BioMed research international.

[18]  Abhijit Gosavi,et al.  Reinforcement Learning: A Tutorial Survey and Recent Advances , 2009, INFORMS J. Comput..

[19]  Chee Kheong Siew,et al.  Can threshold networks be trained directly? , 2006, IEEE Transactions on Circuits and Systems II: Express Briefs.

[20]  Guang-Bin Huang,et al.  Convex incremental extreme learning machine , 2007, Neurocomputing.

[21]  Chee Kheong Siew,et al.  Extreme learning machine: RBF network case , 2004, ICARCV 2004 8th Control, Automation, Robotics and Vision Conference, 2004..

[22]  Lei Chen,et al.  Enhanced random search based incremental extreme learning machine , 2008, Neurocomputing.

[23]  Chee Kheong Siew,et al.  Real-time learning capability of neural networks , 2006, IEEE Trans. Neural Networks.

[24]  Guang-Bin Huang,et al.  Extreme learning machine: a new learning scheme of feedforward neural networks , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[25]  P. Saratchandran,et al.  Multicategory Classification Using An Extreme Learning Machine for Microarray Gene Expression Cancer Diagnosis , 2007, IEEE/ACM Transactions on Computational Biology and Bioinformatics.

[26]  Badong Chen,et al.  Quantized Kernel Recursive Least Squares Algorithm , 2013, IEEE Transactions on Neural Networks and Learning Systems.

[27]  Narasimhan Sundararajan,et al.  Fully complex extreme learning machine , 2005, Neurocomputing.

[28]  Jing Ren,et al.  Modified Newton's method applied to potential field-based navigation for mobile robots , 2006, IEEE Transactions on Robotics.

[29]  Zhiping Lin,et al.  Extreme Learning Machines on High Dimensional and Large Data Applications: A Survey , 2015 .

[30]  María Jesús López Boada,et al.  Visual approach skill for a mobile robot using learning and fusion of simple skills , 2002, Robotics Auton. Syst..

[31]  Silvia Ferrari,et al.  Information-Driven Sensor Path Planning by Approximate Cell Decomposition , 2009, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[32]  Cheng-Wan An,et al.  Mobile robot navigation using neural Q-learning , 2004, Proceedings of 2004 International Conference on Machine Learning and Cybernetics (IEEE Cat. No.04EX826).

[33]  Wei Zheng,et al.  Elman Fuzzy Adaptive Control for Obstacle Avoidance of Mobile Robots Using Hybrid Force/Position Incorporation , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[34]  Chee Kheong Siew,et al.  Universal Approximation using Incremental Constructive Feedforward Networks with Random Hidden Nodes , 2006, IEEE Transactions on Neural Networks.

[35]  Narasimhan Sundararajan,et al.  A Fast and Accurate Online Sequential Learning Algorithm for Feedforward Networks , 2006, IEEE Transactions on Neural Networks.