Learning probabilistic neural representations with randomly connected circuits

Significance We present a theory of neural circuits’ design and function, inspired by the random connectivity of real neural circuits and the mathematical power of random projections. Specifically, we introduce a family of statistical models for large neural population codes, a straightforward neural circuit architecture that would implement these models, and a biologically plausible learning rule for such circuits. The resulting neural architecture suggests a design principle for neural circuit—namely, that they learn to compute the mathematical surprise of their inputs, given past inputs, without an explicit teaching signal. We applied these models to recordings from large neural populations in monkeys’ visual and prefrontal cortices and show them to be highly accurate, efficient, and scalable. The brain represents and reasons probabilistically about complex stimuli and motor actions using a noisy, spike-based neural code. A key building block for such neural computations, as well as the basis for supervised and unsupervised learning, is the ability to estimate the surprise or likelihood of incoming high-dimensional neural activity patterns. Despite progress in statistical modeling of neural responses and deep learning, current approaches either do not scale to large neural populations or cannot be implemented using biologically realistic mechanisms. Inspired by the sparse and random connectivity of real neuronal circuits, we present a model for neural codes that accurately estimates the likelihood of individual spiking patterns and has a straightforward, scalable, efficient, learnable, and realistic neural implementation. This model’s performance on simultaneously recorded spiking activity of >100 neurons in the monkey visual and prefrontal cortices is comparable with or better than that of state-of-the-art models. Importantly, the model can be learned using a small number of samples and using a local learning rule that utilizes noise intrinsic to neural circuits. Slower, structural changes in random connectivity, consistent with rewiring and pruning processes, further improve the efficiency and sparseness of the resulting neural representations. Our results merge insights from neuroanatomy, machine learning, and theoretical neuroscience to suggest random sparse connectivity as a key design principle for neuronal computation.

[1]  József Fiser,et al.  Spontaneous Cortical Activity Reveals Hallmarks of an Optimal Internal Model of the Environment , 2011, Science.

[2]  Yoshua Bengio,et al.  Towards Biologically Plausible Deep Learning , 2015, ArXiv.

[3]  Zeb Kurth-Nelson,et al.  A distributional code for value in dopamine-based reinforcement learning , 2020, Nature.

[4]  G Shahaf,et al.  Learning in Networks of Cortical Neurons , 2001, The Journal of Neuroscience.

[5]  Wulfram Gerstner,et al.  Predicting non-linear dynamics by stable local learning in a recurrent spiking neural network , 2017, eLife.

[6]  Yonina C. Eldar,et al.  Compressed Sensing with Coherent and Redundant Dictionaries , 2010, ArXiv.

[7]  J. Born,et al.  Neuroscience and Biobehavioral Reviews Sleep and Memory in Mammals, Birds and Invertebrates , 2022 .

[8]  Christopher J. Cueva,et al.  Dynamics of Neural Population Responses in Prefrontal Cortex Indicate Changes of Mind on Single Trials , 2014, Current Biology.

[9]  E. Jaynes Information Theory and Statistical Mechanics , 1957 .

[10]  Laurence Aitchison,et al.  The Hamiltonian Brain: Efficient Probabilistic Inference with Excitatory-Inhibitory Neural Circuit Dynamics , 2014, PLoS Comput. Biol..

[11]  Sanjoy Dasgupta,et al.  A neural data structure for novelty detection , 2018, Proceedings of the National Academy of Sciences.

[12]  D. Landau,et al.  Efficient, multiple-range random walk algorithm to calculate the density of states. , 2000, Physical review letters.

[13]  D. Kumaran,et al.  Which computational mechanisms operate in the hippocampus during novelty detection? , 2007, Hippocampus.

[14]  Eero P. Simoncelli,et al.  Spatio-temporal correlations and visual signalling in a complete neuronal population , 2008, Nature.

[15]  Chuan Wang,et al.  Training neural networks with additive noise in the desired signal , 1999, IEEE Trans. Neural Networks.

[16]  Xaq Pitkow,et al.  "Compressive neural representation of sparse, high-dimensional probabilities" , 2012, NIPS.

[17]  Y. Nesterov A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .

[18]  H. Sompolinsky,et al.  Sparseness and Expansion in Sensory Representations , 2014, Neuron.

[19]  Michael J. Berry,et al.  Searching for Collective Behavior in a Large Network of Sensory Neurons , 2013, PLoS Comput. Biol..

[20]  Michael J. Berry,et al.  Weak pairwise correlations imply strongly correlated network states in a neural population , 2005, Nature.

[21]  Nuo Li,et al.  Robust neuronal dynamics in premotor cortex during motor planning , 2016, Nature.

[22]  Jascha Sohl-Dickstein,et al.  A new method for parameter estimation in probabilistic models: Minimum probability flow , 2011, Physical review letters.

[23]  Rob Malouf,et al.  A Comparison of Algorithms for Maximum Entropy Parameter Estimation , 2002, CoNLL.

[24]  Christopher C. Pack,et al.  Bidirectional manipulation of GABAergic inhibition in MT: A comparison of neuronal and psychophysical performance , 2014 .

[25]  Marwan A. Jabri,et al.  Weight perturbation: an optimal architecture and learning technique for analog VLSI feedforward and recurrent multilayer networks , 1992, IEEE Trans. Neural Networks.

[26]  Lloyd H. Michael,et al.  The Guide for the Care and Use of Laboratory Animals. , 2016, ILAR journal.

[27]  Rishidev Chaudhuri,et al.  Bipartite expander Hopfield networks as self-decoding high-capacity error correcting codes , 2019, NeurIPS.

[28]  S. Leibler,et al.  Neuronal couplings between retinal ganglion cells inferred by efficient inverse statistical physics methods , 2009, Proceedings of the National Academy of Sciences.

[29]  Takaki Komiyama,et al.  Subtype-specific plasticity of inhibitory circuits in motor cortex during motor learning , 2015, Nature Neuroscience.

[30]  Jonathon Shlens,et al.  The Structure of Multi-Neuron Firing Patterns in Primate Retina , 2006, The Journal of Neuroscience.

[31]  H. Sompolinsky,et al.  Compressed sensing, sparsity, and dimensionality in neuronal information processing and data analysis. , 2012, Annual review of neuroscience.

[32]  W. Bialek,et al.  A sensory source for motor variation , 2005, Nature.

[33]  Eero P. Simoncelli,et al.  Partitioning neuronal variability , 2014, Nature Neuroscience.

[34]  Emery N. Brown,et al.  Dynamic Analysis of Neural Encoding by Point Process Adaptive Filtering , 2004, Neural Computation.

[35]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.

[36]  Bruno A. Olshausen,et al.  Modeling Higher-Order Correlations within Cortical Microcolumns , 2014, PLoS Comput. Biol..

[37]  Ann M Hermundstad,et al.  Adaptive coding for dynamic sensory inference , 2017, bioRxiv.

[38]  G. Turrigiano Too many cooks? Intrinsic and synaptic homeostatic mechanisms in cortical circuit refinement. , 2011, Annual review of neuroscience.

[39]  J. Movshon,et al.  The analysis of visual motion: a comparison of neuronal and psychophysical performance , 1992, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[40]  Alexandre Pouget,et al.  Probabilistic Interpretation of Population Codes , 1996, Neural Computation.

[41]  T. Sejnowski,et al.  Reliability of spike timing in neocortical neurons. , 1995, Science.

[42]  M. Tsodyks,et al.  Synaptic Theory of Working Memory , 2008, Science.

[43]  J. DiCarlo,et al.  Using goal-driven deep learning models to understand sensory cortex , 2016, Nature Neuroscience.

[44]  A. Faisal,et al.  Noise in the nervous system , 2008, Nature Reviews Neuroscience.

[45]  Mark Shein-Idelson,et al.  Slow waves, sharp waves, ripples, and REM in sleeping dragons , 2016, Science.

[46]  A. Barto,et al.  Novelty or Surprise? , 2013, Front. Psychol..

[47]  S. Denéve,et al.  Neural processing as causal inference , 2011, Current Opinion in Neurobiology.

[48]  Dmitri B. Chklovskii,et al.  A Normative Theory of Adaptive Dimensionality Reduction in Neural Networks , 2015, NIPS.

[49]  József Fiser,et al.  Neural Variability and Sampling-Based Probabilistic Representations in the Visual Cortex , 2016, Neuron.

[50]  L. Pinneo On noise in the nervous system. , 1966, Psychological review.

[51]  Stefano Fusi,et al.  The Sparseness of Mixed Selectivity Neurons Controls the Generalization–Discrimination Trade-Off , 2013, The Journal of Neuroscience.

[52]  John M. Beggs,et al.  A Maximum Entropy Model Applied to Spatial and Temporal Correlations from Cortical Networks In Vitro , 2008, The Journal of Neuroscience.

[53]  W. Senn,et al.  Learning by the Dendritic Prediction of Somatic Spiking , 2014, Neuron.

[54]  Peter Dayan,et al.  A Neural Substrate of Prediction and Reward , 1997, Science.

[55]  Robert Gütig,et al.  Spiking neurons can discover predictive features by aggregate-label learning , 2016, Science.

[56]  Haim Sompolinsky,et al.  Optimal Degrees of Synaptic Connectivity , 2017, Neuron.

[57]  H. Seung,et al.  Learning in Spiking Neural Networks by Reinforcement of Stochastic Synaptic Transmission , 2003, Neuron.

[58]  Rajesh P. N. Rao,et al.  Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. , 1999 .

[59]  Xiaohui Xie,et al.  Learning in neural networks by reinforcement of irregular spiking. , 2004, Physical review. E, Statistical, nonlinear, and soft matter physics.

[60]  Naoshige Uchida,et al.  Demixed principal component analysis of neural population data , 2014, eLife.

[61]  Maneesh Sahani,et al.  Flexible and accurate inference and learning for deep generative models , 2018, NeurIPS.

[62]  Division on Earth Guide for the Care and Use of Laboratory Animals , 1996 .

[63]  R. Segev,et al.  Sparse low-order interaction network underlies a highly correlated and learnable neural population code , 2011, Proceedings of the National Academy of Sciences.

[64]  L. F. Abbott,et al.  Generating Coherent Patterns of Activity from Chaotic Neural Networks , 2009, Neuron.

[65]  H. Markram,et al.  The neural code between neocortical pyramidal neurons depends on neurotransmitter release probability. , 1997, Proceedings of the National Academy of Sciences of the United States of America.

[66]  Movshon J. Anthony Visual response properties of V1 neurons projecting to V2 in macaque , 2009 .

[67]  J. Csicsvari,et al.  Organization of cell assemblies in the hippocampus , 2003, Nature.

[68]  L. Abbott,et al.  Random Convergence of Olfactory Inputs in the Drosophila Mushroom Body , 2013, Nature.

[69]  Jonathan D. Cohen,et al.  Computational roles for dopamine in behavioural control , 2004, Nature.

[70]  G. Palm Evidence, information, and surprise , 1981, Biological Cybernetics.

[71]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[72]  G. Rainer,et al.  Cognitive neuroscience: Neural mechanisms for detecting and remembering novel events , 2003, Nature Reviews Neuroscience.

[73]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.