Neural Modularity Helps Organisms Evolve to Learn New Skills without Forgetting Old Skills

A long-standing goal in artificial intelligence is creating agents that can learn a variety of different skills for different problems. In the artificial intelligence subfield of neural networks, a barrier to that goal is that when agents learn a new skill they typically do so by losing previously acquired skills, a problem called catastrophic forgetting. That occurs because, to learn the new task, neural learning algorithms change connections that encode previously acquired skills. How networks are organized critically affects their learning dynamics. In this paper, we test whether catastrophic forgetting can be reduced by evolving modular neural networks. Modularity intuitively should reduce learning interference between tasks by separating functionality into physically distinct modules in which learning can be selectively turned on or off. Modularity can further improve learning by having a reinforcement learning module separate from sensory processing modules, allowing learning to happen only in response to a positive or negative reward. In this paper, learning takes place via neuromodulation, which allows agents to selectively change the rate of learning for each neural connection based on environmental stimuli (e.g. to alter learning in specific locations based on the task at hand). To produce modularity, we evolve neural networks with a cost for neural connections. We show that this connection cost technique causes modularity, confirming a previous result, and that such sparsely connected, modular networks have higher overall performance because they learn new skills faster while retaining old skills more and because they have a separate reinforcement learning module. Our results suggest (1) that encouraging modularity in neural networks may help us overcome the long-standing barrier of networks that cannot learn new skills without forgetting old ones, and (2) that one benefit of the modularity ubiquitous in the brains of natural animals might be to alleviate the problem of catastrophic forgetting.

[1]  B. Underwood,et al.  Fate of first-list associations in transfer theory. , 1959, Journal of experimental psychology.

[2]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[3]  Geoffrey E. Hinton,et al.  How Learning Can Guide Evolution , 1996, Complex Syst..

[4]  Michael McCloskey,et al.  Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .

[5]  R Ratcliff,et al.  Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. , 1990, Psychological review.

[6]  Bernard Widrow,et al.  30 years of adaptive neural networks: perceptron, Madaline, and backpropagation , 1990, Proc. IEEE.

[7]  Robert M. French,et al.  Using Semi-Distributed Representations to Overcome Catastrophic Forgetting in Connectionist Networks , 1991 .

[8]  E. Capaldi,et al.  The organization of behavior. , 1992, Journal of applied behavior analysis.

[9]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[10]  Jeffrey L. Elman,et al.  Learning and Evolution in Neural Networks , 1994, Adapt. Behav..

[11]  S. Lewandowsky,et al.  Catastrophic interference in neural networks , 1995 .

[12]  James L. McClelland,et al.  Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. , 1995, Psychological review.

[13]  Anthony V. Robins,et al.  Catastrophic Forgetting, Rehearsal and Pseudorehearsal , 1995, Connect. Sci..

[14]  S.J.J. Smith,et al.  Empirical Methods for Artificial Intelligence , 1995 .

[15]  Keiji Tanaka,et al.  Inferotemporal cortex and object vision. , 1996, Annual review of neuroscience.

[16]  Giles Mayley,et al.  Landscapes, Learning Costs, and Genetic Assimilation , 1996, Evolutionary Computation.

[17]  Robert M. French,et al.  Pseudo-recurrent Connectionist Networks: An Approach to the 'Sensitivity-Stability' Dilemma , 1997, Connect. Sci..

[18]  V. Mountcastle The columnar organization of the neocortex. , 1997, Brain : a journal of neurology.

[19]  R. French Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.

[20]  Takahiro Sasaki,et al.  Evolving Learnable Neural Networks Under Changing Environments with Various Rates of Inheritance of Acquired Characters: Comparison of Darwinian and Lamarckian Evolution , 1999, Artificial Life.

[21]  X. Yao Evolving Artificial Neural Networks , 1999 .

[22]  Xin Yao,et al.  Evolving artificial neural networks , 1999, Proc. IEEE.

[23]  Bernard Ans,et al.  Neural networks with a self-refreshing memory: Knowledge transfer in sequential learning tasks without catastrophic forgetting , 2000, Connect. Sci..

[24]  Dario Floreano,et al.  Evolution of Adaptive Synapses: Robots with Fast Adaptive Behavior in New Environments , 2001, Evolutionary Computation.

[25]  S. Carroll Chance and necessity: the evolution of morphological complexity and diversity , 2001, Nature.

[26]  B. Burrell,et al.  Learning in simple systems , 2001, Current Opinion in Neurobiology.

[27]  Isaac Meilijson,et al.  Evolution of Reinforcement Learning in Uncertain Environments: A Simple Explanation for Complex Foraging Behaviors , 2002, Adapt. Behav..

[28]  Kalyanmoy Deb,et al.  A fast and elitist multiobjective genetic algorithm: NSGA-II , 2002, IEEE Trans. Evol. Comput..

[29]  Dario Floreano,et al.  Levels of dynamics and adaptive behavior in evolutionary neural controllers , 2002 .

[30]  Risto Miikkulainen,et al.  Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.

[31]  Risto Miikkulainen,et al.  Active Guidance for a Finless Rocket Using Neuroevolution , 2003, GECCO.

[32]  Risto Miikkulainen,et al.  A Taxonomy for Artificial Embryogeny , 2003, Artificial Life.

[33]  John R. Koza,et al.  Genetic Programming IV: Routine Human-Competitive Machine Intelligence , 2003 .

[34]  T. Jay Dopamine: a potential substrate for synaptic plasticity and memory mechanisms , 2003, Progress in Neurobiology.

[35]  Raul Rodriguez-Esteban,et al.  Global optimization of cerebral cortex layout. , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[36]  L. Abbott,et al.  Synaptic computation , 2004, Nature.

[37]  R. Guimerà,et al.  Modularity from fluctuations in random graphs and complex networks. , 2004, Physical review. E, Statistical, nonlinear, and soft matter physics.

[38]  Ronald,et al.  Learning representations by backpropagating errors , 2004 .

[39]  L’oubli catastrophique it,et al.  Avoiding catastrophic forgetting by coupling two reverberating neural networks , 2004 .

[40]  Masahiro Fujita,et al.  Autonomous evolution of dynamic gaits with two quadruped robots , 2005, IEEE Transactions on Robotics.

[41]  John A. Bullinaria,et al.  The Evolution of Minimal Catastrophic Forgetting in Neural Systems , 2005 .

[42]  G. Striedter Principles of brain evolution. , 2005 .

[43]  U. Alon,et al.  Spontaneous evolution of modularity and network motifs. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[44]  M E J Newman,et al.  Modularity and community structure in networks. , 2006, Proceedings of the National Academy of Sciences of the United States of America.

[45]  Uri Alon,et al.  An Introduction to Systems Biology , 2006 .

[46]  Yong-Yeol Ahn,et al.  Wiring cost in the organization of a biological neuronal network , 2005, q-bio/0505009.

[47]  G. Wagner,et al.  The road to modularity , 2007, Nature Reviews Genetics.

[48]  Hod Lipson,et al.  Principles of modularity, regularity, and hierarchy for scalable systems , 2007 .

[49]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[50]  Andrea Soltoggio Neural Plasticity and Minimal Topologies for Reward-Based Learning , 2008, 2008 Eighth International Conference on Hybrid Intelligent Systems.

[51]  M. Pigliucci Is evolvability evolvable? , 2008, Nature Reviews Genetics.

[52]  Arend Hintze,et al.  Evolution of Complex Modular Biological Networks , 2007, PLoS Comput. Biol..

[53]  Dario Floreano,et al.  Evolutionary Advantages of Neuromodulated Plasticity in Dynamic, Reward-based Scenarios , 2008, ALIFE.

[54]  E A Leicht,et al.  Community structure in directed networks. , 2007, Physical review letters.

[55]  Ulrik Brandes,et al.  On Modularity Clustering , 2008, IEEE Transactions on Knowledge and Data Engineering.

[56]  Charles E. Hughes,et al.  How novelty search escapes the deceptive trap of learning to learn , 2009, GECCO.

[57]  Charles Ofria,et al.  Evolving coordinated quadruped gaits with the HyperNEAT generative encoding , 2009, 2009 IEEE Congress on Evolutionary Computation.

[58]  D. Stephens,et al.  Components of change in the evolution of learning and unlearned preference , 2009, Proceedings of the Royal Society B: Biological Sciences.

[59]  Santo Fortunato,et al.  Community detection in graphs , 2009, ArXiv.

[60]  Anthony Kulis,et al.  Bio-Inspired Artificial Intelligence: Theories, Methods, and Technologies , 2009, Scalable Comput. Pract. Exp..

[61]  V. Ramakrishnan,et al.  Measurement of the top-quark mass with dilepton events selected using neuroevolution at CDF. , 2008, Physical review letters.

[62]  Andreas Wagner,et al.  Specialization Can Drive the Evolution of Modularity , 2010, PLoS Comput. Biol..

[63]  Stéphane Doncieux,et al.  Sferesv2: Evolvin' in the multi-core world , 2010, IEEE Congress on Evolutionary Computation.

[64]  Simon Haykin,et al.  Neural Networks and Learning Machines , 2010 .

[65]  Charles Ofria,et al.  Investigating whether hyperNEAT produces modular neural networks , 2010, GECCO '10.

[66]  Kenneth O. Stanley,et al.  Constraining connectivity to encourage modularity in HyperNEAT , 2011, GECCO '11.

[67]  Kenneth O. Stanley,et al.  On the Performance of Indirect Encoding Across the Continuum of Regularity , 2011, IEEE Transactions on Evolutionary Computation.

[68]  Jason D. Lohn,et al.  Computer-Automated Evolution of an X-Band Antenna for NASA's Space Technology 5 Mission , 2011, Evolutionary Computation.

[69]  Robert Anemone,et al.  Finding fossils in new ways: An artificial neural network approach to predicting the location of productive fossil localities , 2011, Evolutionary anthropology.

[70]  Jean-Baptiste Mouret,et al.  On the relationships between synaptic plasticity and generative systems , 2011, GECCO '11.

[71]  Boye Annfelt Høverstad,et al.  Noise and the Evolution of Neural Network Modularity , 2011, Artificial Life.

[72]  Stéphane Doncieux,et al.  Encouraging Behavioral Diversity in Evolutionary Robotics: An Empirical Study , 2012, Evolutionary Computation.

[73]  E. Bizzi,et al.  A theory for how sensorimotor skills are learned and retained in noisy and nonstationary neural circuits , 2013, Proceedings of the National Academy of Sciences.

[74]  Kai Olav Ellefsen Balancing the Costs and Benefits of Learning Ability , 2013, ECAL.

[75]  Martial Mermillod,et al.  The stability-plasticity dilemma: investigating the continuum from catastrophic forgetting to age-limited learning effects , 2013, Front. Psychol..

[76]  Jean-Baptiste Mouret,et al.  On the Relationships between Generative Encodings, Regularity, and Learning Abilities when Evolving Plastic Artificial Neural Networks , 2013, PloS one.

[77]  Geoffrey J. Gordon,et al.  Artificial Intelligence in Medicine: 17th Conference on Artificial Intelligence in Medicine, AIME 2019, Poznan, Poland, June 26–29, 2019, Proceedings , 2019, Lecture Notes in Computer Science.

[78]  Hod Lipson,et al.  The evolutionary origins of modularity , 2012, Proceedings of the Royal Society B: Biological Sciences.