Dimensionality Reduction on Spatio-Temporal Maximum Entropy Models of Spiking Networks

Maximum entropy models (MEM) have been widely used in the last 10 years to characterize the statistics of networks of spiking neurons. A major drawback of this approach is that the number of parameters used in the statistical model increases very fast with the network size, hindering its interpretation and fast computation. Here, we present a novel framework of dimensionality reduction for generalized MEM handling spatio-temporal correlations. This formalism is based on information geometry where a MEM is a point on a large-dimensional manifold. We exploit the geometrical properties of this manifold in order to find a projection on a lower dimensional space that best captures the high-order statistics. This allows us to define a quantitative criterion that we call the “degree of compressibility” of the neuronal code. A powerful aspect of this method is that it does not require fitting the model. Indeed, the matrix defining the metric of the manifold is computed directly via the data without parameters fitting. The method is first validated using synthetic data generated by a known statistics. We then analyze a MEM having more parameters than the underlying data statistics and show that our method detects the extra dimensions. We then test it on experimental retinal data. We record retinal ganglion cells (RGC) spiking data using multi-electrode arrays (MEA) under different visual stimuli: spontaneous activity, white noise stimulus, and natural scene. Using our method, we report a dimensionality reduction up to 50% for retinal data. As we show, this is quite a huge reduction compared to a randomly generated spike train, suggesting that the neuronal code, in these experiments, is highly compressible. This additionally shows that the dimensionality reduction depends on the stimuli statistics, supporting the idea that sensory networks adapt to stimuli statistics by modifying the level of redundancy. Author Summary Maximum entropy models (MEM) have been widely used to characterize the statistics of networks of spiking neurons. However, as the network size increases, the number of model parameters increases rapidly, hindering its interpretation and fast computation. Here, we propose a method to evaluate the dimensionality reduction of MEM, based on the geometrical properties of the manifold best capturing the network high-order statistics. Our method is validated with synthetic data using independent or correlated neural responses. Importantly, we show that dimensionality reduction depends on the stimuli statistics, supporting the idea that sensory networks adapt to stimuli statistics modifying the level of redundancy.

[1]  Bruno Cessac,et al.  Dynamics and spike trains statistics in conductance-based Integrate-and-Fire neural networks with chemical and electric synapses , 2012, BMC Neuroscience.

[2]  M. Laubach,et al.  Redundancy and Synergy of Neuronal Ensembles in Motor Cortex , 2005, The Journal of Neuroscience.

[3]  Bruno Cessac,et al.  Spatio-temporal spike train analysis for large scale networks using the maximum entropy principle and Monte Carlo method , 2012, 1209.3886.

[4]  Tim Gollisch,et al.  Eye Smarter than Scientists Believed: Neural Computations in Circuits of the Retina , 2010, Neuron.

[5]  Michael J. Berry,et al.  Gibbs distribution analysis of temporal correlations structure in retina ganglion cells , 2011, Journal of Physiology - Paris.

[6]  R. Monasson,et al.  Small-correlation expansions for the inverse Ising problem , 2008, 0811.3574.

[7]  I. Mastromatteo,et al.  On the criticality of inferred models , 2011, 1102.1624.

[8]  Jonathon Shlens,et al.  The Structure of Multi-Neuron Firing Patterns in Primate Retina , 2006, The Journal of Neuroscience.

[9]  DANNY CALEGARI,et al.  Thermodynamic Formalism , 2021, Lecture Notes in Mathematics.

[10]  Eero P. Simoncelli,et al.  Spatio-temporal correlations and visual signalling in a complete neuronal population , 2008, Nature.

[11]  David J Schwab,et al.  Nonlinear dendritic integration of electrical and chemical synaptic inputs drives fine-scale correlations , 2014, Nature Neuroscience.

[12]  William Bialek,et al.  Spikes: Exploring the Neural Code , 1996 .

[13]  Brent Doiron,et al.  Theory of oscillatory firing induced by spatially correlated noise and delayed inhibitory feedback. , 2005, Physical review. E, Statistical, nonlinear, and soft matter physics.

[14]  Bruno Cessac,et al.  Spike Train Statistics from Empirical Facts to Theory: The Case of the Retina , 2013 .

[15]  Gasper Tkacik,et al.  Optimal population coding by noisy spiking neurons , 2010, Proceedings of the National Academy of Sciences.

[16]  R. Segev,et al.  The Architecture of Functional Interaction Networks in the Retina , 2011, The Journal of Neuroscience.

[17]  B Cessac,et al.  Exact computation of the maximum-entropy potential of spiking neural-network models. , 2014, Physical review. E, Statistical, nonlinear, and soft matter physics.

[18]  R. Segev,et al.  Sparse low-order interaction network underlies a highly correlated and learnable neural population code , 2011, Proceedings of the National Academy of Sciences.

[19]  M. Fisher,et al.  Universal critical amplitudes in finite-size scaling , 1984 .

[20]  Eero P. Simoncelli,et al.  Modeling the Impact of Common Noise Inputs on the Network Activity of Retinal Ganglion Cells Action Editor: Brent Doiron , 2022 .

[21]  Pierre Kornprobst,et al.  PRANAS: A New Platform for Retinal Analysis and Simulation , 2017, Front. Neuroinform..

[22]  Michael J. Berry,et al.  Weak pairwise correlations imply strongly correlated network states in a neural population , 2005, Nature.

[23]  Michael J. Berry,et al.  Redundancy in the Population Code of the Retina , 2005, Neuron.

[24]  A. Daffertshofer,et al.  Functional Connectivity Dynamics of the Resting State across the Human Adult Lifespan , 2017, bioRxiv.

[25]  Bruno Cessac,et al.  Parameter Estimation for Spatio-Temporal Maximum Entropy Distributions: Application to Neural Spike Trains , 2014, Entropy.

[26]  Shun-ichi Amari,et al.  Methods of information geometry , 2000 .

[27]  G. Keller Equilibrium States in Ergodic Theory , 1998 .

[28]  Luca Berdondini,et al.  Sloppiness in Spontaneously Active Neuronal Networks , 2015, The Journal of Neuroscience.

[29]  Sami El Boustani,et al.  Prediction of spatiotemporal patterns of neural activity from pairwise correlations. , 2009, Physical review letters.

[30]  Hans-Otto Georgii,et al.  Gibbs Measures and Phase Transitions , 1988 .

[31]  H. Bozdogan Model selection and Akaike's Information Criterion (AIC): The general theory and its analytical extensions , 1987 .

[32]  Vijay Balasubramanian,et al.  Transformation of Stimulus Correlations by the Retina , 2013, PLoS Comput. Biol..

[33]  Eric Shea-Brown,et al.  Impact of Network Structure and Cellular Response on Spike Time Correlations , 2011, PLoS Comput. Biol..

[34]  J. B. Demb,et al.  Contrast Adaptation in Subthreshold and Spiking Responses of Mammalian Y-Type Retinal Ganglion Cells , 2005, The Journal of Neuroscience.

[35]  Michael J. Berry,et al.  The simplest maximum entropy model for collective behavior in a neural network , 2012, 1207.6319.

[36]  E. Jaynes Information Theory and Statistical Mechanics , 1957 .

[37]  Ute Dreher,et al.  Statistical Mechanics Rigorous Results , 2016 .

[38]  Michael J. Berry,et al.  Recording spikes from a large fraction of the ganglion cells in a retinal patch , 2004, Nature Neuroscience.

[39]  Rava Azeredo da Silveira,et al.  High-Fidelity Coding with Correlated Neurons , 2013, PLoS Comput. Biol..

[40]  Y. Sinai GIBBS MEASURES IN ERGODIC THEORY , 1972 .

[41]  Vijay Balasubramanian,et al.  Statistical Inference, Occam's Razor, and Statistical Mechanics on the Space of Probability Distributions , 1996, Neural Computation.

[42]  Gašper Tkačik,et al.  Maximum entropy models as a tool for building precise neural controls , 2017, Current Opinion in Neurobiology.

[43]  A. Palacios,et al.  Role of connexin channels in the retinal light response of a diurnal rodent , 2014, Front. Cell. Neurosci..

[44]  Nicolas Brunel,et al.  Fast Global Oscillations in Networks of Integrate-and-Fire Neurons with Low Firing Rates , 1999, Neural Computation.