Mitigating Catastrophic Forgetting with Complementary Layered Learning

Catastrophic forgetting is a stability–plasticity imbalance that causes a machine learner to lose previously gained knowledge that is critical for performing a task. The imbalance occurs in transfer learning, negatively affecting the learner’s performance, particularly in neural networks and layered learning. This work proposes a complementary learning technique that introduces long- and short-term memory to layered learning to reduce the negative effects of catastrophic forgetting. In particular, this work proposes the dual memory system in the non-neural network approaches of evolutionary computation and Q-learning instances of layered learning because these techniques are used to develop decision-making capabilities for physical robots. Experiments evaluate the new learning augmentation in a multi-agent system simulation, where autonomous unmanned aerial vehicles learn to collaborate and maneuver to survey an area effectively. Through these direct-policy and value-based learning experiments, the proposed complementary layered learning is demonstrated to significantly improve task performance over standard layered learning, successfully balancing stability and plasticity.

[1]  E. Callaway What's next for AlphaFold and the AI protein-folding revolution , 2022, Nature.

[2]  Md. Saiful Islam,et al.  Foreign Exchange Currency Rate Prediction using a GRU-LSTM Hybrid Network , 2020 .

[3]  Wojciech M. Czarnecki,et al.  Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.

[4]  Yoshua Bengio,et al.  Light Gated Recurrent Units for Speech Recognition , 2018, IEEE Transactions on Emerging Topics in Computational Intelligence.

[5]  Motonobu Hattori,et al.  Avoiding Catastrophic Forgetting by a Dual-Network Memory Model Using a Chaotic Neural Network , 2009 .

[6]  Frank Dellaert,et al.  RoboCup 2007: Robot Soccer World Cup XI, July 9-10, 2007, Atlanta, GA, USA , 2008, RoboCup.

[7]  Ingo Wegener,et al.  On the analysis of a simple evolutionary algorithm on quadratic pseudo-boolean functions , 2005, J. Discrete Algorithms.

[8]  Steven M. Gustafson,et al.  Layered Learning in Genetic Programming for a Cooperative Robot Soccer Problem , 2001, EuroGP.

[9]  Jürgen Schmidhuber,et al.  Learning to Forget: Continual Prediction with LSTM , 2000, Neural Computation.

[10]  Anthony V. Robins,et al.  Catastrophic Forgetting and the Pseudorehearsal Solution in Hopfield-type Networks , 1998, Connect. Sci..

[11]  Manuela M. Veloso,et al.  Layered Approach to Learning Client Behaviors in the Robocup Soccer Server , 1998, Appl. Artif. Intell..

[12]  Robert M. French,et al.  Pseudo-recurrent Connectionist Networks: An Approach to the 'Sensitivity-Stability' Dilemma , 1997, Connect. Sci..

[13]  S. Hochreiter,et al.  Long Short-Term Memory , 1997, Neural Computation.

[14]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[15]  Anton V. Eremeev,et al.  A Study on Performance of the (1+1)-Evolutionary Algorithm , 2002, FOGA.

[16]  Michael McCloskey,et al.  Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .