Exploration of genetic network programming with two-stage reinforcement learning for mobile robot
暂无分享,去创建一个
Kotaro Hirasawa | Hsien-I Lin | A. N. Afandi | Ilham A. E. Zaeni | Yogi Dwi Mahandi | Siti Sendari | K. Hirasawa | Hsien-I Lin | A. Afandi | S. Sendari | I. Zaeni
[1] Shingo Mabu,et al. A Graph-Based Evolutionary Algorithm: Genetic Network Programming (GNP) and Its Extension Using Reinforcement Learning , 2007, Evolutionary Computation.
[2] Kotaro Hirasawa,et al. Comparison between Genetic Network Programming (GNP) and Genetic Programming (GP) , 2001, Proceedings of the 2001 Congress on Evolutionary Computation (IEEE Cat. No.01TH8546).
[3] Marco A. Contreras-Cruz,et al. Mobile robot path planning using artificial bee colony and evolutionary programming , 2015, Appl. Soft Comput..
[4] Lazhar Khriji,et al. Mobile Robot Navigation Based on Q-Learning Technique , 2011 .
[5] Kun Deng,et al. Balancing exploration and exploitation: a new algorithm for active machine learning , 2005, Fifth IEEE International Conference on Data Mining (ICDM'05).
[6] Leslie Pack Kaelbling,et al. Effective reinforcement learning for mobile robots , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).
[7] 木村 正行,et al. Robustness of robot programs generated by genetic programming , 1996 .
[8] Xin Yao,et al. Evolving artificial neural networks , 1999, Proc. IEEE.
[9] Eduardo F. Morales,et al. A Two-Stage Relational Reinforcement Learning with Continuous Actions for Real Service Robots , 2009, MICAI.
[10] Kazushi Murakoshi,et al. A parameter control method inspired from neuromodulators in reinforcement learning , 2003, Proceedings 2003 IEEE International Symposium on Computational Intelligence in Robotics and Automation. Computational Intelligence in Robotics and Automation for the New Millennium (Cat. No.03EX694).
[11] Xuesong Yan. An Improved Robot Path Planning Algorithm , 2012 .
[12] Argel A. Bandala,et al. Path planning for mobile robots using genetic algorithm and probabilistic roadmap , 2017, 2017IEEE 9th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment and Management (HNICEM).
[13] Sunil Kumar Kashyap,et al. Matrix-Binary Codes based Genetic Algorithm for path planning of mobile robot , 2017, Comput. Electr. Eng..
[14] Mohd Faisal Ibrahim,et al. Ga-based optimisation of a lidar feedback autonomous mobile robot navigation system , 2018 .
[15] Junichiro Yoshimoto,et al. Control of exploitation-exploration meta-parameter in reinforcement learning , 2002, Neural Networks.
[16] Shingo Mabu,et al. Fuzzy Genetic Network Programming with Noises for Mobile Robot Navigation , 2011, J. Adv. Comput. Intell. Intell. Informatics.
[17] Shingo Mabu,et al. Genetic Network Programming with Reinforcement Learning and Its Application to Making Mobile Robot Behavior , 2006 .
[18] John R. Koza,et al. Genetic Programming II , 1992 .
[19] Hirokazu Watabe,et al. Automatic generation of behaviors for mobile robot by GA with automatically generated action rule-base , 2000, 2000 26th Annual Conference of the IEEE Industrial Electronics Society. IECON 2000. 2000 IEEE International Conference on Industrial Electronics, Control and Instrumentation. 21st Century Technologies.
[20] Meng Cai,et al. Path planning for unmanned aerial vehicles based on genetic programming , 2016, CCDC 2016.
[21] Daoyi Dong,et al. Robust Quantum-Inspired Reinforcement Learning for Robot Navigation , 2012, IEEE/ASME Transactions on Mechatronics.
[22] S. Mabu,et al. Two-Stage Reinforcement Learning based on Genetic Network Programming for mobile robot , 2012, 2012 Proceedings of SICE Annual Conference (SICE).
[23] Shan Liang,et al. Optimal Path Planning for Mobile Robot Using Tailored Genetic Algorithm , 2014 .
[24] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[25] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..
[26] Michel Tokic. Adaptive ε-greedy Exploration in Reinforcement Learning Based on Value Differences , 2010 .