Energy Efficient Neurons With Generalized Inverse Gaussian Conditional and Marginal Hitting Times

Neuronal information processing is energetically costly. Energy supply restrictions on information processing have resulted in the evolution of brains to compute and communicate information with remarkable energy efficiency. Indeed, energy minimization subject to functional constraints is a unifying principle. Toward better comprehension of neuronal information processing and communication from an information-energy standpoint, we consider a continuous time, continuous state-space neuron model with a generalized inverse Gaussian (GIG) conditional density. This GIG model arises from a Levy diffusion that contains both homogeneous Poisson processes and Wiener processes with drift as special cases. We show that, when the energy constraints consist of a tripartite family apropos of the GIG model, the distribution of input excitation, $\Lambda $ , that maximizes bits per Joule (bpJ) generates an output interspike interval duration $T$ that possesses a related GIG marginal distribution. Most importantly, we obtain the exact expression for the bpJ-maximizing distribution of $\Lambda $ in terms of Bessel functions that are readily evaluated numerically. This permits display of the strictly concave curves of bpJ-maximized mutual information $I(\Lambda ;T)$ . Because bpJ equals bits per second/watt, these curves serve as the Shannon-theory capacity v. power curves for GIG neuron models. We show that the noise in GIG channels is essentially multiplicative, so our results may have significance for broader classes of fading channels.

[1]  Jeffrey S. Diamond,et al.  Asynchronous release of synaptic vesicles determines the time course of the AMPA receptor-mediated EPSC , 1995, Neuron.

[2]  William B Levy,et al.  Energy-Efficient Neuronal Computation via Quantal Synaptic Failures , 2002, The Journal of Neuroscience.

[3]  W. McCulloch,et al.  The limiting information capacity of a neuronal link , 1952 .

[4]  L. Aiello,et al.  The Expensive-Tissue Hypothesis: The Brain and the Digestive System in Human and Primate Evolution , 1995, Current Anthropology.

[5]  Toby Berger,et al.  A Mathematical Theory of Energy Efficient Neural Computation and Communication , 2010, IEEE Transactions on Information Theory.

[6]  S. Laughlin,et al.  Fly Photoreceptors Demonstrate Energy-Information Trade-Offs in Neural Coding , 2007, PLoS biology.

[7]  Toby Berger,et al.  A Berger-Levy energy efficient neuron model with unequal synaptic weights , 2012, 2012 IEEE International Symposium on Information Theory Proceedings.

[8]  T. Sejnowski,et al.  Synchrony of Thalamocortical Inputs Maximizes Cortical Reliability , 2010, Science.

[9]  Jonathon Howard,et al.  Drawing an elephant with four complex parameters , 2010 .

[10]  Toby Berger,et al.  Energy Efficient Neurons With Generalized Inverse Gaussian Conditional and Marginal Hitting Times , 2015, IEEE Transactions on Information Theory.

[11]  J. Nadal,et al.  What can we learn from synaptic weight distributions? , 2007, Trends in Neurosciences.

[12]  Emery N. Brown,et al.  Computational Neuroscience: A Comprehensive Approach , 2022 .

[13]  D. Attwell,et al.  Synaptic Energy Use and Supply , 2012, Neuron.

[14]  G. Mitchison Neuronal branching patterns and the economy of cortical wiring , 1991, Proceedings of the Royal Society of London. Series B: Biological Sciences.

[15]  T. Sejnowski,et al.  Correlated neuronal activity and the flow of neural information , 2001, Nature Reviews Neuroscience.

[16]  B. Mandelbrot,et al.  RANDOM WALK MODELS FOR THE SPIKE ACTIVITY OF A SINGLE NEURON. , 1964, Biophysical journal.

[17]  Ole E. Barndorff-Nielsen,et al.  Hyperbolic Distributions and Ramifications: Contributions to Theory and Application , 1981 .

[18]  William B. Levy,et al.  Neural computation at the thermal limit , 2014 .

[19]  T. Sejnowski,et al.  Metabolic cost as a unifying principle governing neuronal biophysics , 2010, Proceedings of the National Academy of Sciences.

[20]  Simon B. Laughlin,et al.  Action Potential Energy Efficiency Varies Among Neuron Types in Vertebrates and Invertebrates , 2010, PLoS Comput. Biol..

[21]  Rob R. de Ruyter van Steveninck,et al.  The metabolic cost of neural information , 1998, Nature Neuroscience.

[22]  Michael J. Berry,et al.  Metabolically Efficient Information Processing , 2001, Neural Computation.

[23]  William B. Levy,et al.  Energy Efficient Neural Codes , 1996, Neural Computation.

[24]  William Bialek,et al.  Reading a Neural Code , 1991, NIPS.

[25]  Ole E. Barndorff-Nielsen,et al.  First hitting time models for the generalized inverse Gaussian distribution , 1978 .

[26]  Toby Berger,et al.  Energy efficient neurons with generalized inverse Gaussian interspike interval durations , 2011, 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[27]  B. Willmore,et al.  Sparse coding in striate and extrastriate visual cortex. , 2011, Journal of neurophysiology.

[28]  Terrence J Sejnowski,et al.  Communication in Neuronal Networks , 2003, Science.

[29]  Warren S. McCulloch,et al.  An upper bound on the informational capacity of a synapse , 1952, ACM '52.

[30]  Satish Iyengar,et al.  Modeling neural activity using the generalized inverse Gaussian distribution , 1997, Biological Cybernetics.

[31]  David Attwell,et al.  The Energetics of CNS White Matter , 2012, The Journal of Neuroscience.

[32]  Freeman Dyson,et al.  A meeting with Enrico Fermi , 2004, Nature.

[33]  Sergio Verdú,et al.  Spectral efficiency in the wideband regime , 2002, IEEE Trans. Inf. Theory.

[34]  S. Laughlin,et al.  An Energy Budget for Signaling in the Grey Matter of the Brain , 2001, Journal of cerebral blood flow and metabolism : official journal of the International Society of Cerebral Blood Flow and Metabolism.