A self-organizing short-term dynamical memory network

Working memory requires information about external stimuli to be represented in the brain even after those stimuli go away. This information is encoded in the activities of neurons, and neural activities change over timescales of tens of milliseconds. Information in working memory, however, is retained for tens of seconds, suggesting the question of how time-varying neural activities maintain stable representations. Prior work shows that, if the neural dynamics are in the 'null space' of the representation - so that changes to neural activity do not affect the downstream read-out of stimulus information - then information can be retained for periods much longer than the time-scale of individual-neuronal activities. The prior work, however, requires precisely constructed synaptic connectivity matrices, without explaining how this would arise in a biological neural network. To identify mechanisms through which biological networks can self-organize to learn memory function, we derived biologically plausible synaptic plasticity rules that dynamically modify the connectivity matrix to enable information storing. Networks implementing this plasticity rule can successfully learn to form memory representations even if only 10% of the synapses are plastic, they are robust to synaptic noise, and they can represent information about multiple stimuli.

[1]  Sophie Denève,et al.  Enforcing balance allows local supervised learning in spiking recurrent networks , 2015, NIPS.

[2]  Nicolas Brunel,et al.  Dynamics of Networks of Excitatory and Inhibitory Neurons in Response to Time-Dependent Inputs , 2011, Front. Comput. Neurosci..

[3]  A. Compte,et al.  Bump attractor dynamics in prefrontal cortex explains behavioral precision in spatial working memory , 2014, Nature Neuroscience.

[4]  P. Goldman-Rakic,et al.  Prefrontal neuronal activity in rhesus monkeys performing a delayed anti-saccade task , 1993, Nature.

[5]  R. Romo,et al.  Timing and neural encoding of somatosensory parametric working memory in macaque prefrontal cortex. , 2003, Cerebral cortex.

[6]  Miguel C. Soriano Viewpoint: Reservoir Computing Speeds Up , 2017 .

[7]  Masanobu Inubushi,et al.  Reservoir Computing Beyond Memory-Nonlinearity Trade-off , 2017, Scientific Reports.

[8]  Daniel Cownden,et al.  Random feedback weights support learning in deep neural networks , 2014, ArXiv.

[9]  Michael Robert DeWeese,et al.  A Sparse Coding Model with Synaptically Local Plasticity and Spiking Neurons Can Account for the Diverse Shapes of V1 Simple Cell Receptive Fields , 2011, PLoS Comput. Biol..

[10]  Su Z. Hong,et al.  Distinct Eligibility Traces for LTP and LTD in Cortical Synapses , 2015, Neuron.

[11]  Wieland Brendel,et al.  Unsupervised learning of an efficient short-term memory network , 2014, NIPS.

[12]  Earl K. Miller,et al.  Limber Neurons for a Nimble Mind , 2013, Neuron.

[13]  Christian K. Machens,et al.  Learning optimal spike-based representations , 2012, NIPS.

[14]  R. Romo,et al.  Neuronal correlates of parametric working memory in the prefrontal cortex , 1999, Nature.

[15]  Sophie Denève,et al.  The Brain as an Efficient and Robust Adaptive Learner , 2017, Neuron.

[16]  N. Cowan,et al.  The Magical Mystery Four , 2010, Current directions in psychological science.

[17]  Benjamin Schrauwen,et al.  Information Processing Capacity of Dynamical Systems , 2012, Scientific Reports.

[18]  A. Peters,et al.  Synapses are lost during aging in the primate prefrontal cortex , 2008, Neuroscience.

[19]  M. Carandini Amplification of Trial-to-Trial Response Variability by Neurons in Visual Cortex , 2004, PLoS biology.

[20]  C. Constantinidis,et al.  The neuroscience of working memory capacity and training , 2016, Nature Reviews Neuroscience.

[21]  Christos Constantinidis,et al.  Stable population coding for working memory coexists with heterogeneous neural dynamics in prefrontal cortex , 2016, Proceedings of the National Academy of Sciences.

[22]  Timothy P Lillicrap,et al.  Towards deep learning with segregated dendrites , 2016, eLife.

[23]  Sarah Marzen,et al.  The difference between memory and prediction in linear recurrent networks , 2017, Physical review. E.

[24]  Ranulfo Romo,et al.  Basic mechanisms for graded persistent activity: discrete attractors, continuous attractors, and dynamic representations , 2003, Current Opinion in Neurobiology.

[25]  Nuo Li,et al.  Robust neuronal dynamics in premotor cortex during motor planning , 2016, Nature.

[26]  Dmitri B. Chklovskii,et al.  Neuronal Circuits Underlying Persistent Representations Despite Time Varying Activity , 2012, Current Biology.

[27]  J. Zylberberg,et al.  Mechanisms of Persistent Activity in Cortical Circuits: Possible Neural Substrates for Working Memory. , 2017, Annual review of neuroscience.

[28]  Zhengya Zhang,et al.  A Sparse Coding Neural Network ASIC With On-Chip Learning for Feature Extraction and Encoding , 2015, IEEE Journal of Solid-State Circuits.

[29]  Bernhard Hellwig,et al.  A quantitative analysis of the local connectivity between pyramidal neurons in layers 2/3 of the rat visual cortex , 2000, Biological Cybernetics.

[30]  C. Meunier,et al.  Modulation of Synaptic Plasticity in the Cortex Needs to Understand All the Players , 2017, Front. Synaptic Neurosci..

[31]  G. E. Alexander,et al.  Neuron Activity Related to Short-Term Memory , 1971, Science.

[32]  L. Abbott,et al.  Beyond the edge of chaos: amplification and temporal integration by recurrent networks in the chaotic regime. , 2011, Physical review. E, Statistical, nonlinear, and soft matter physics.

[33]  Xiaohui Xie,et al.  Spike-based Learning Rules and Stabilization of Persistent Neural Activity , 1999, NIPS.