Structured Mean-field Variational Inference and Learning in Winner-take-all Spiking Neural Networks

The Bayesian view of the brain hypothesizes that the brain constructs a generative model of the world, and uses it to make inferences via Bayes' rule. Although many types of approximate inference schemes have been proposed for hierarchical Bayesian models of the brain, the questions of how these distinct inference procedures can be realized by hierarchical networks of spiking neurons remains largely unresolved. Based on a previously proposed multi-compartment neuron model in which dendrites perform logarithmic compression, and stochastic spiking winner-take-all (WTA) circuits in which firing probability of each neuron is normalized by activities of other neurons, here we construct Spiking Neural Networks that perform \emph{structured} mean-field variational inference and learning, on hierarchical directed probabilistic graphical models with discrete random variables. In these models, we do away with symmetric synaptic weights previously assumed for \emph{unstructured} mean-field variational inference by learning both the feedback and feedforward weights separately. The resulting online learning rules take the form of an error-modulated local Spike-Timing-Dependent Plasticity rule. Importantly, we consider two types of WTA circuits in which only one neuron is allowed to fire at a time (hard WTA) or neurons can fire independently (soft WTA), which makes neurons in these circuits operate in regimes of temporal and rate coding respectively. We show how the hard WTA circuits can be used to perform Gibbs sampling whereas the soft WTA circuits can be used to implement a message passing algorithm that computes the marginals approximately. Notably, a simple change in the amount of lateral inhibition realizes switching between the hard and soft WTA spiking regimes. Hence the proposed network provides a unified view of the two previously disparate modes of inference and coding by spiking neurons.

[1]  Karol Gregor,et al.  Neural Variational Inference and Learning in Belief Networks , 2014, ICML.

[2]  Wolfgang Maass,et al.  Bayesian Computation Emerges in Generic Cortical Microcircuits through Spike-Timing-Dependent Plasticity , 2013, PLoS Comput. Biol..

[3]  Britt Anderson,et al.  Conditional probability modulates visual search efficiency , 2013, Front. Hum. Neurosci..

[4]  Tai Sing Lee,et al.  Hierarchical Bayesian inference in the visual cortex. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[5]  Stefan Habenschuss,et al.  Homeostatic plasticity in Bayesian spiking networks as Expectation Maximization with posterior constraints , 2012, NIPS.

[6]  J. F. Kolen,et al.  Backpropagation without weight transport , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).

[7]  Yoshua Bengio,et al.  Dendritic cortical microcircuits approximate the backpropagation algorithm , 2018, NeurIPS.

[8]  William T. Freeman,et al.  Understanding belief propagation and its generalizations , 2003 .

[9]  Geoffrey E. Hinton,et al.  The Helmholtz Machine , 1995, Neural Computation.

[10]  Michael I. Jordan,et al.  Graphical Models, Exponential Families, and Variational Inference , 2008, Found. Trends Mach. Learn..

[11]  V. Borkar,et al.  Stochastic approximation , 2013, Resonance.

[12]  F. Gabbiani,et al.  Logarithmic Compression of Sensory Signals within the Dendritic Tree of a Collision-Sensitive Neuron , 2012, The Journal of Neuroscience.

[13]  Hedvig Kjellström,et al.  Advances in Variational Inference , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  Joydeep Ghosh,et al.  The pi-sigma network: an efficient higher-order neural network for pattern classification and function approximation , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[15]  David M. Blei,et al.  Variational Inference: A Review for Statisticians , 2016, ArXiv.

[16]  Subutai Ahmad,et al.  Why Neurons Have Thousands of Synapses, a Theory of Sequence Memory in Neocortex , 2015, Front. Neural Circuits.

[17]  Hideaki Shimazaki,et al.  The principles of adaptation in organisms and machines I: machine learning, information theory, and thermodynamics , 2019, 1902.11233.

[18]  Matthias S. Keil,et al.  Dendritic Pooling of Noisy Threshold Processes Can Explain Many Properties of a Collision-Sensitive Visual Neuron , 2015, PLoS Comput. Biol..

[19]  Peter C. Humphreys,et al.  Deep Learning without Weight Transport , 2019, NeurIPS.

[20]  Lin Wu,et al.  What-and-Where to Match: Deep Spatially Multiplicative Integration Networks for Person Re-identification , 2017, Pattern Recognit..

[21]  Martin J. Wainwright,et al.  Tree-based reparameterization framework for analysis of sum-product and related algorithms , 2003, IEEE Trans. Inf. Theory.

[22]  Chris Dyer,et al.  Neural Arithmetic Logic Units , 2018, NeurIPS.

[23]  Rajesh P. N. Rao Hierarchical Bayesian Inference in Networks of Spiking Neurons , 2004, NIPS.

[24]  Paul Miller,et al.  Spiking neuron network Helmholtz machine , 2015, Front. Comput. Neurosci..

[25]  Andreas Voss,et al.  Components of task switching: a closer look at task switching and cue switching. , 2014, Acta psychologica.

[26]  Gerhard Lakemeyer,et al.  Exploring artificial intelligence in the new millennium , 2003 .

[27]  D. Knill,et al.  The Bayesian brain: the role of uncertainty in neural coding and computation , 2004, Trends in Neurosciences.

[28]  Tiejun Huang,et al.  Winner-Take-All as Basic Probabilistic Inference Unit of Neuronal Circuits , 2018, ArXiv.

[29]  Karl J. Friston The free-energy principle: a unified brain theory? , 2010, Nature Reviews Neuroscience.

[30]  Rajesh P. N. Rao,et al.  Bayesian brain : probabilistic approaches to neural coding , 2006 .

[31]  David Kappel,et al.  STDP Installs in Winner-Take-All Circuits an Online Approximation to Hidden Markov Model Learning , 2014, PLoS Comput. Biol..

[32]  Wolfgang Maass,et al.  Probabilistic Inference in General Graphical Models through Sampling in Stochastic Networks of Spiking Neurons , 2011, PLoS Comput. Biol..

[33]  Xaq Pitkow,et al.  Inference by Reparameterization in Neural Population Codes , 2016, NIPS.

[34]  Karl J. Friston,et al.  Neuronal message passing using Mean-field, Bethe, and Marginal approximations , 2019, Scientific Reports.

[35]  Kendra S. Burbank Mirrored STDP Implements Autoencoder Learning in a Network of Spiking Neurons , 2015, PLoS Comput. Biol..

[36]  Chong Wang,et al.  Stochastic variational inference , 2012, J. Mach. Learn. Res..

[37]  Michael Schmitt,et al.  On the Complexity of Computing and Learning with Multiplicative Neural Networks , 2002, Neural Computation.

[38]  Wolfgang Maass,et al.  Learning Probabilistic Inference through Spike-Timing-Dependent Plasticity123 , 2016, eNeuro.

[39]  Xiaolin Hu,et al.  Hierarchical Bayesian Inference and Learning in Spiking Neural Networks , 2019, IEEE Transactions on Cybernetics.