To learn useful dynamics on long time scales, neurons must use plasticity rules that account for long-term, circuit-wide effects of synaptic changes. In other words, neural circuits must solve a credit assignment problem to appropriately assign responsibility for global network behavior to individual circuit components. Furthermore, biological constraints demand that plasticity rules are spatially and temporally local; that is, synaptic changes can depend only on variables accessible to the pre- and postsynaptic neurons. While artificial intelligence offers a computational solution for credit assignment, namely backpropagation through time (BPTT), this solution is wildly biologically implausible. It requires both nonlocal computations and unlimited memory capacity, as any synaptic change is a complicated function of the entire history of network activity. Similar nonlocality issues plague other approaches such as FORCE (Sussillo et al. 2009). Overall, we are still missing a model for learning in recurrent circuits that both works computationally and uses only local updates. Leveraging recent advances in machine learning on approximating gradients for BPTT, we derive biologically plausible plasticity rules that enable recurrent networks to accurately learn long-term dependencies in sequential data. The solution takes the form of neurons with segregated voltage compartments, with several synaptic sub-populations that have different functional properties. The network operates in distinct phases during which each synaptic sub-population is updated by its own local plasticity rule. Our results provide new insights into the potential roles of segregated dendritic compartments, branch-specific inhibition, and global circuit phases in learning.
[1]
Alex Graves,et al.
Decoupled Neural Interfaces using Synthetic Gradients
,
2016,
ICML.
[2]
Thomas Klausberger,et al.
Temporal redistribution of inhibition over neuronal subcellular domains underlies state-dependent rhythmic change of excitability in the hippocampus
,
2014,
Philosophical Transactions of the Royal Society B: Biological Sciences.
[3]
L. F. Abbott,et al.
Generating Coherent Patterns of Activity from Chaotic Neural Networks
,
2009,
Neuron.
[4]
Razvan Pascanu,et al.
On the difficulty of training recurrent neural networks
,
2012,
ICML.
[5]
Timothy P Lillicrap,et al.
Towards deep learning with segregated dendrites
,
2016,
eLife.
[6]
S. Siegelbaum,et al.
A Role for Synaptic Inputs at Distal Dendrites: Instructive Signals for Hippocampal Long-Term Plasticity
,
2007,
Neuron.