Hardware design of autonomous snake-like robot for reinforcement learning based on environment

In this paper, we propose the design of a robot with a snake-like body based on a test environment. We explore the abstraction of state-action spaces for reinforcement learning. Additionally, we discuss the versatility of the proposed mechanism by showing that different tasks can be completed by simply changing the reward of the reinforcement learning. Finally, we mention the importance of a body design based on an environment by considering the concept of ecological niches.

[1]  Kazuyuki Ito,et al.  Extended QDSEGA for controlling real robots - acquisition of locomotion patterns for snake-like robot , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[2]  Aristidis Likas Reinforcement Learning Using the Stochastic Fuzzy Min–Max Neural Network , 2004, Neural Processing Letters.

[3]  Shigenobu Kobayashi,et al.  Reinforcement learning of walking behavior for a four-legged robot , 2001, Proceedings of the 40th IEEE Conference on Decision and Control (Cat. No.01CH37228).

[4]  Dongbing Gu,et al.  Reinforcement learning of fuzzy logic controllers for quadruped walking robots , 2002 .

[5]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[6]  Tetsushi Kamegawa,et al.  Development of the snake-like rescue robot "kohga" , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[7]  K. Ito,et al.  Autonomous control of real snake-like robot using reinforcement learning; Abstraction of state-action space using properties of real world , 2007, 2007 3rd International Conference on Intelligent Sensors, Sensor Networks and Information.

[8]  Howie Choset,et al.  A mobile hyper redundant mechanism for search and rescue tasks , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[9]  Minoru Asada,et al.  Behavior coordination for a mobile robot using modular reinforcement learning , 1996, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems. IROS '96.

[10]  Thomas Christaller,et al.  A robot snake to inspect broken buildings , 2000, Proceedings. 2000 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2000) (Cat. No.00CH37113).

[11]  Charles W. Anderson,et al.  Reinforcement Learning with Modular Neural Networks for Control , 1994 .

[12]  Richard S. Sutton,et al.  Introduction to Reinforcement Learning , 1998 .

[13]  Fumitoshi Matsuno,et al.  Development of the Snake-like Rescue Robot. , 2004 .

[14]  Kazuyuki Ito,et al.  Autonomous control of a snake-like robot utilizing passive mechanism , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[15]  Masayuki Arai,et al.  Development of "Souryu-III": connected crawler vehicle for inspection inside narrow and winding spaces , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[16]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[17]  Kenji Doya,et al.  Neural mechanisms of learning and control , 2001 .

[18]  Shigeo Hirose,et al.  Stabilization of the head of an undulating snake-like robot , 2007, 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[19]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.