Training a Network of Spiking Neurons with Equilibrium Propagation

Backpropagation is almost universally used to train artificial neural networks. However, there are several reasons that backpropagation could not be plausibly implemented by biological neurons. Among these are the facts that (1) biological neurons appear to lack any mechanism for sending gradients backwards across synapses, and (2) biological “spiking” neurons emit binary signals, whereas back-propagation requires that neurons communicate real numbers between one another. Recently (Scellier and Bengio, 2017), demonstrated an alternative to backpropagation, called Equilibrium Propagation, wherein gradients are implicitly computed by the dynamics of the neural network, so that neurons do not need an internal mechanism for backpropagation of gradients. This provides an interesting solution to problem (1). In this paper, we address problem (2) by proposing a way in which Equilibrium Propagation can be implemented with neurons which are constrained to just communicate binary values at each time step. We show that with appropriate step-size annealing, we can converge to the same fixed-point as a real-valued neural network, and that with predictive coding, we can make this convergence much faster. We demonstrate that the resulting model can be used to train a neural network using the update scheme from Equilibrium propagation.

[1]  Young C. Yoon,et al.  LIF and Simplified SRM Neurons Encode Signals Into Spikes via a Form of Asynchronous Pulse Sigma–Delta Modulation , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Guillaume Charpiat,et al.  Training recurrent networks online without backtracking , 2015, ArXiv.

[3]  Gabor C. Temes,et al.  Oversampling delta-sigma data converters : theory, design, and simulation , 1992 .

[4]  Yoshua Bengio,et al.  Extending the Framework of Equilibrium Propagation to General Dynamics , 2018, ICLR.

[5]  Daniel Cownden,et al.  Random feedback weights support learning in deep neural networks , 2014, ArXiv.

[6]  Yoshua Bengio,et al.  Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation , 2016, Front. Comput. Neurosci..

[7]  M. Fu,et al.  An Overview of Stochastic Approximation , 2015 .

[8]  M. N. S. Swamy Oversampled , Linear Predictive and Noise-Shaping Coders of Order N > 1 , 2017 .

[9]  Warren B. Powell,et al.  Adaptive stepsizes for recursive estimation with applications in approximate dynamic programming , 2006, Machine Learning.

[10]  Francis Crick,et al.  The recent excitement about neural networks , 1989, Nature.

[11]  Jonghan Shin,et al.  Adaptive noise shaping neural spike encoding and decoding , 2001, Neurocomputing.

[12]  J J Hopfield,et al.  Neurons with graded response have collective computational properties like those of two-state neurons. , 1984, Proceedings of the National Academy of Sciences of the United States of America.

[13]  H. Robbins A Stochastic Approximation Method , 1951 .

[14]  Sander M. Bohte,et al.  Fast and Efficient Asynchronous Neural Computation with Adapting Spiking Neural Networks , 2016, ArXiv.

[15]  S. Laughlin,et al.  Predictive coding: a fresh view of inhibition in the retina , 1982, Proceedings of the Royal Society of London. Series B. Biological Sciences.

[16]  Yoshua Bengio,et al.  STDP as presynaptic activity times rate of change of postsynaptic activity , 2015, 1509.05936.

[17]  Max Welling,et al.  Temporally Efficient Deep Learning with Spikes , 2018, ICLR.

[18]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[19]  Yann Ollivier,et al.  Unbiased Online Recurrent Optimization , 2017, ICLR.

[20]  Dmitri B. Chklovskii,et al.  Automatic Adaptation to Fast Input Changes in a Time-Invariant Neural Circuit , 2015, PLoS Comput. Biol..