Reinforcement learning based group navigation approach for multiple autonomous robotic system

In several complex applications, the use of multiple autonomous robotic systems (ARS) becomes necessary to achieve different tasks such as foraging and transport of heavy and large objects with less cost and more efficiency. They have to achieve a high level of flexibility, adaptability and efficiency in real environments. In this paper, a reinforcement learning (RL) based group navigation approach for multiple ARS is suggested. Indeed, the robots must have the ability to form geometric figures and navigate without collisions while maintaining the formation. Thus, each robot must learn how to take its place in the formation and avoid obstacles and other ARS from its interaction with the environment. This approach must provide ARS with capability to acquire the group navigation approach among several ARS from elementary behaviors by learning with trial and error search. Then, simulation results display the ability of the suggested approach to provide ARS with capability to navigate in a group formation in dynamic environments.

[1]  Sandra Clara Gadanho,et al.  Reinforcement learning in autonomous robots : an empirical investigation of the role of emotions , 1999 .

[2]  Maja J. Matarić,et al.  Designing emergent behaviors: from local interactions to collective intelligence , 1993 .

[3]  A. Damasio Descartes’ Error. Emotion, Reason and the Human Brain. New York (Grosset/Putnam) 1994. , 1994 .

[4]  Maja J. Mataric,et al.  A general algorithm for robot formations using local sensing and minimal communication , 2002, IEEE Trans. Robotics Autom..

[5]  Ichiro Suzuki,et al.  Distributed algorithms for formation of geometric patterns with many mobile robots , 1996, J. Field Robotics.

[6]  Yoshikazu Arai,et al.  Multilayered reinforcement learning for complicated collision avoidance problems , 1998, Proceedings. 1998 IEEE International Conference on Robotics and Automation (Cat. No.98CH36146).

[7]  Ming Tan,et al.  Multi-Agent Reinforcement Learning: Independent versus Cooperative Agents , 1997, ICML.

[8]  A. Damasio Descartes' error: emotion, reason, and the human brain. avon books , 1994 .

[9]  Stefan Schaal,et al.  Learning Robot Control , 2002 .

[10]  Claude F. Touzet,et al.  Neural reinforcement learning for behaviour synthesis , 1997, Robotics Auton. Syst..

[11]  Yoshikazu Arai,et al.  Collision avoidance among multiple autonomous mobile robots using LOCISS (locally communicable infrared sensory system) , 1996, Proceedings of IEEE International Conference on Robotics and Automation.

[12]  Vladimir Polotski,et al.  Decentralized Control of Two Cooperative Car-Like Robots Performing a Transportation Task , 1998 .

[13]  O. Azouaoui,et al.  Evolution, Behavior, and Intelligence of Autonomous Robotic Systems (ARS) , 1998 .

[14]  Maja J. Mataric,et al.  Robot formations using only local sensing and control , 2001, Proceedings 2001 IEEE International Symposium on Computational Intelligence in Robotics and Automation (Cat. No.01EX515).

[15]  J. Y. S. Luh,et al.  Coordination and control of a group of small mobile robots , 1994, Proceedings of the 1994 IEEE International Conference on Robotics and Automation.

[16]  Steven Douglas Whitehead,et al.  Reinforcement learning for the adaptive control of perception and action , 1992 .

[17]  Karsten Berns,et al.  Reinforcement-learning For The Control Of An Autonomous Mobile Robot , 1992, Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems.

[18]  Satoshi Fujita,et al.  Learning-based automatic generation of collision avoidance algorithms for multiple autonomous mobile robots , 1998, Proceedings. 1998 IEEE/RSJ International Conference on Intelligent Robots and Systems. Innovations in Theory, Practice and Applications (Cat. No.98CH36190).

[19]  Rolf Pfeifer,et al.  Understanding intelligence , 2020, Inequality by Design.

[20]  Maja J. Mataric,et al.  Interaction and intelligent behavior , 1994 .

[21]  Yoshikazu Arai,et al.  Adaptive behavior acquisition of collision avoidance among multiple autonomous mobile robots , 1997, Proceedings of the 1997 IEEE/RSJ International Conference on Intelligent Robot and Systems. Innovative Robotics for Real-World Applications. IROS '97.

[22]  Leslie Pack Kaelbling,et al.  Learning in embedded systems , 1993 .

[23]  Sebastian Thrun,et al.  Issues in Using Function Approximation for Reinforcement Learning , 1999 .

[24]  Maja J. Matarić,et al.  Robots in Formation Using Local Information , 2002 .

[25]  Chris Watkins,et al.  Learning from delayed rewards , 1989 .

[26]  Yoshio Kawauchi,et al.  A principle of distributed decision making of Cellular Robotic System (CEBOT) , 1993, [1993] Proceedings IEEE International Conference on Robotics and Automation.

[27]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[28]  Minoru Asada,et al.  Modular Learning System and Scheduling for Behavior Acquisition in Multi-agent Environment , 2004, RoboCup.

[29]  Sebastian Thrun,et al.  A lifelong learning perspective for mobile robot control , 1994, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS'94).

[30]  Paul Keng-Chieh Wang Navigation strategies for multiple autonomous mobile robots moving in formation , 1991, J. Field Robotics.

[31]  Yoshikazu Arai,et al.  Realization of autonomous navigation in multirobot environment , 1998, Proceedings. 1998 IEEE/RSJ International Conference on Intelligent Robots and Systems. Innovations in Theory, Practice and Applications (Cat. No.98CH36190).

[32]  Toru Ishida,et al.  Towards organizational problem solving , 1993, [1993] Proceedings IEEE International Conference on Robotics and Automation.