Learning of action patterns and reactive behavior plans via a novel two-layered ethology-based action selection mechanism

The two most important abilities for a robot to survive in a given environment are selecting and learning the most appropriate actions in a given situation. Historically, they have also been the biggest problems in robotics. To help solve this problem, we propose a two-layered action selection mechanism (ASM) which designates an action pattern layer and a reactive behavior plan layer. In the reactive behavior plan layer, a task is selected by comparing behavior motivation values that, in an animal, correspond to external stimuli as well as internal states due to hormones. After a task is selected, its corresponding reactive behavior plan is executed as a set of sequential dynamic behavior motivations (DBMs), each of which is associated with an action pattern. In the action pattern layer, each action pattern can be functionally decomposed into primitive motor actions. Shortest path-based Q-learning (SPQL) is incorporated into both the reactive behavior plan and action pattern layers. In the reactive behavior plan layer, relationships between perceptions and action patterns are learned to satisfy a given motivation, as well as the relative priorities of these relationships. In the action pattern layer, the relations between sensory states and primitive motor actions can be learned. To establish the validity of our proposed ASM, experiments with our real designed robot was illustrated together with simulations.

[1]  Il Hong Suh,et al.  A reinforcement learning approach involving a shortest path finding algorithm , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[2]  Il Hong Suh,et al.  Design and implementation of a behavior-based control and learning architecture for mobile robots , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[3]  William Rowan,et al.  The Study of Instinct , 1953 .

[4]  Rodney A. Brooks,et al.  A Robust Layered Control Syste For A Mobile Robot , 2022 .

[5]  Pattie Maes,et al.  A bottom-up mechanism for behavior selection in an artificial creature , 1991 .

[7]  J. K. Rosenblatt,et al.  A fine-grained alternative to the subsumption architecture for mobile robot control , 1989, International 1989 Joint Conference on Neural Networks.

[8]  M. O. Foltermann,et al.  Intelligent design. , 2007, The Pharos of Alpha Omega Alpha-Honor Medical Society. Alpha Omega Alpha.

[9]  L W White Old tricks for new dogs. , 1996, Journal of clinical orthodontics : JCO.

[10]  Stewart W. Wilson,et al.  A Finer-Grained Motivational Model of Behaviour Sequencing , 1996 .

[11]  Mark Humphreys,et al.  Action selection methods using reinforcement learning , 1997 .

[12]  Emmet Spier,et al.  A Finer-Grained Motivational Model of Behaviour Sequencing , 1996 .

[13]  Nando de Freitas,et al.  Sequential Monte Carlo in Practice , 2001 .

[14]  P. Maes,et al.  Old tricks, new dogs: ethology and interactive creatures , 1997 .

[15]  H. Evans The Study of Instinct , 1952 .

[16]  Paolo Pirjanian An Overview of System Architectures for Action Selection in Mobile Robotics , 1997 .

[17]  Marco Colombetti,et al.  Robot Shaping: Developing Autonomous Agents Through Learning , 1994, Artif. Intell..

[18]  I. Suh,et al.  A Novel Action Selection Mechanism for Intelligent Service Robots , 2003 .

[19]  Christopher Mark Witkowski,et al.  Schemes for learning and behaviour : a new expectancy model , 2013 .