Working memory facilitates reward-modulated Hebbian learning in recurrent neural networks

Reservoir computing is a powerful tool to explain how the brain learns temporal sequences, such as movements, but existing learning schemes are either biologically implausible or too inefficient to explain animal performance. We show that a network can learn complicated sequences with a reward-modulated Hebbian learning rule if the network of reservoir neurons is combined with a second network that serves as a dynamic working memory and provides a spatio-temporal backbone signal to the reservoir. In combination with the working memory, reward-modulated Hebbian learning of the readout neurons performs as well as FORCE learning, but with the advantage of a biologically plausible interpretation of both the learning rule and the learning paradigm.

[1]  L. Christophorou Science , 2018, Emerging Dynamics: Science, Energy, Society and Values.

[2]  October I Physical Review Letters , 2022 .

[3]  E. G. Jones Cerebral Cortex , 1987, Cerebral Cortex.

[4]  M. V. Rossum,et al.  In Neural Computation , 2022 .

[5]  D. Wilkin,et al.  Neuron , 2001, Brain Research.

[6]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[7]  Front , 2020, 2020 Fourth World Conference on Smart Trends in Systems, Security and Sustainability (WorldS4).