R-HybrID: Evolution of Agent Controllers with a Hybrisation of Indirect and Direct Encodings

Neuroevolution, the optimisation of artificial neural networks (ANNs) through evolutionary computation, is a promising approach to the synthesis of controllers for autonomous agents. Traditional neuroevolution approaches employ direct encodings, which are limited in their ability to evolve complex or large-scale controllers because each ANN parameter is independently optimised. Indirect encodings, on the other hand, facilitate scalability because each gene can be reused multiple times to construct the ANN, but are biased towards regularity and can become ineffective when irregularity is required. To address such limitations, we introduce a novel algorithm called R-HybrID. In R-HybrID, controllers have both indirectly encoded and directly encoded structure. Because the portion of structure following a specific encoding is under evolutionary control, R-HybrID can automatically find an appropriate encoding combination for a given task. We assess the performance of R-HybrID in three tasks: (i) a high-dimensional visual discrimination task that requires geometric principles to be evolved, (ii) a challenging benchmark for modular robotics, and (iii) a memory task that has proven difficult for current algorithms because it requires effectively accumulating neural structure for cognitive behaviour to emerge. Our results show that R-HybrID consistently outperforms three state-of-the-art neuroevolution algorithms, and effectively evolves complex controllers and behaviours.

[1]  Gary J. Koehler,et al.  New stopping criterion for genetic algorithms , 2000, Eur. J. Oper. Res..

[2]  Risto Miikkulainen,et al.  A Taxonomy for Artificial Embryogeny , 2003, Artificial Life.

[3]  Thomas Schmickl,et al.  Coupled inverted pendulums: a benchmark for evolving decentral controllers in modular robotics , 2011, GECCO '11.

[4]  Anders Lyhne Christensen,et al.  odNEAT: An Algorithm for Distributed Online, Onboard Evolution of Robot Behaviours , 2012, ALIFE.

[5]  Inman Harvey,et al.  Evolutionary Robotics: A Survey of Applications and Problems , 1998, EvoRobot.

[6]  Gregory J. Barlow,et al.  Article in Press Robotics and Autonomous Systems ( ) – Robotics and Autonomous Systems Fitness Functions in Evolutionary Robotics: a Survey and Analysis , 2022 .

[7]  Peter J. Bentley,et al.  Three Ways to Grow Designs: A Comparison of Evolved Embryogenies for a Design Problem , 1999 .

[8]  Stéphane Doncieux,et al.  With a little help from selection pressures: evolution of memory in robot controllers , 2012, ALIFE.

[9]  Kenneth O. Stanley,et al.  HyperNEAT: The First Five Years , 2014, Growing Adaptive Machines.

[10]  M. Kendall Statistical Methods for Research Workers , 1937, Nature.

[11]  Francesco Mondada,et al.  The e-puck, a Robot Designed for Education in Engineering , 2009 .

[12]  Risto Miikkulainen,et al.  Efficient evolution of neural networks through complexification , 2004 .

[13]  Kenneth O. Stanley,et al.  On the Performance of Indirect Encoding Across the Continuum of Regularity , 2011, IEEE Transactions on Evolutionary Computation.

[14]  G. Miller,et al.  Artificial Evolution: A New Path for Artificial Intelligence? , 1997, Brain and Cognition.

[15]  Anders Lyhne Christensen,et al.  Online Evolution of Adaptive Robot Behaviour , 2014, Int. J. Nat. Comput. Res..

[16]  Kenneth O. Stanley,et al.  A Case Study on the Critical Role of Geometric Regularity in Machine Learning , 2008, AAAI.

[17]  Risto Miikkulainen,et al.  Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.

[18]  Charles Ofria,et al.  Evolving coordinated quadruped gaits with the HyperNEAT generative encoding , 2009, 2009 IEEE Congress on Evolutionary Computation.

[19]  Anders Lyhne Christensen,et al.  Adaptation of Robot Behaviour through Online Evolution and Neuromodulated Learning , 2012, IBERAMIA.

[20]  Kenneth O. Stanley,et al.  Compositional Pattern Producing Networks : A Novel Abstraction of Development , 2007 .

[21]  Charles E. Hughes,et al.  Evolving plastic neural networks with novelty search , 2010, Adapt. Behav..

[22]  S. Gould The Structure of Evolutionary Theory , 2002 .

[23]  Kenneth O. Stanley,et al.  A Hypercube-Based Encoding for Evolving Large-Scale Neural Networks , 2009, Artificial Life.

[24]  Joel Lehman,et al.  Evolving policy geometry for scalable multiagent learning , 2010, AAMAS.

[25]  Y. Niv,et al.  Evolution of Reinforcement Learning in Uncertain Environments: A Simple Explanation for Complex Foraging Behaviors , 2002 .

[26]  Ming Li,et al.  An Introduction to Kolmogorov Complexity and Its Applications , 1997, Texts in Computer Science.

[27]  Hod Lipson,et al.  Principles of modularity, regularity, and hierarchy for scalable systems , 2007 .

[28]  Heiko Hamann,et al.  Impact of neuron models and network structure on evolving modular robot neural network controllers , 2012, GECCO '12.

[29]  Anders Lyhne Christensen,et al.  odNEAT: An Algorithm for Decentralised Online Evolution of Robotic Controllers , 2015, Evolutionary Computation.

[30]  Risto Miikkulainen,et al.  Effective diversity maintenance in deceptive domains , 2013, GECCO '13.

[31]  Geoffrey E. Hinton,et al.  How Learning Can Guide Evolution , 1996, Complex Syst..

[32]  Kenneth O. Stanley,et al.  Abandoning Objectives: Evolution Through the Search for Novelty Alone , 2011, Evolutionary Computation.

[33]  Stéphane Doncieux,et al.  Encouraging Behavioral Diversity in Evolutionary Robotics: An Empirical Study , 2012, Evolutionary Computation.

[34]  Francesco Mondada,et al.  Automatic creation of an autonomous agent: genetic evolution of a neural-network driven robot , 1994 .

[35]  Dario Floreano,et al.  Neuroevolution: from architectures to learning , 2008, Evol. Intell..

[36]  D. Olton Mazes, maps, and memory. , 1979, The American psychologist.

[37]  Joel Lehman,et al.  Overcoming deception in evolution of cognitive behaviors , 2014, GECCO.

[38]  Sebastian Risi,et al.  Indirectly Encoding Neural Plasticity as a Pattern of Local Rules , 2010, SAB.