Automata Computation with Hodgkin-Huxley Based Neural Networks Composed of Synfire Rings

Recent results have shown that finite state automata can be simulated by recurrent neural networks composed of synfire rings. The simulation process was shown to work correctly in the cases of Boolean neural networks and of Izhikevich spiking neural networks. In this paper, we generalize these results to the very biological context of the Hodgkin-Huxley neural model. We prove that any finite state automaton can be simulated by a Hodgkin-Huxley based recurrent neural network composed of synfire rings. In this framework, the inhibitory system ensuring the transition between the successive rings can be significantly simplified. These results show that a neuro-inspired paradigm of abstract computation based on sustained activities of neural assemblies is indeed possible, and potentially harnessable. They also constitute a first step towards the implementation of biological neural computers.

[1]  Noga Alon,et al.  Efficient simulation of finite automata by neural nets , 1991, JACM.

[2]  Giovanni Soda,et al.  Unified Integration of Explicit Knowledge and Learning by Example in Recurrent Networks , 1995, IEEE Trans. Knowl. Data Eng..

[3]  Alessandro E. P. Villa,et al.  Recurrent Neural Networks and Super-Turing Interactive Computation , 2015 .

[4]  W. Pitts,et al.  A Logical Calculus of the Ideas Immanent in Nervous Activity (1943) , 2021, Ideas That Created the Future.

[5]  C. Lee Giles,et al.  Constructing deterministic finite-state automata in recurrent neural networks , 1996, JACM.

[6]  Heikki Hyotyniemi,et al.  Turing Machines Are Recurrent Neural Networks , 1996 .

[7]  P. Frasconi,et al.  Representation of Finite State Automata in Recurrent Radial Basis Function Networks , 1996, Machine Learning.

[8]  Alessandro E. P. Villa,et al.  The expressive power of analog recurrent neural networks on infinite input streams , 2012, Theor. Comput. Sci..

[9]  Marvin Minsky,et al.  Computation : finite and infinite machines , 2016 .

[10]  江河 Time is Precious , 2003 .

[11]  Don R. Hush,et al.  Bounds on the complexity of recurrent neural network implementations of finite state machines , 1993, Neural Networks.

[12]  C. Lee Giles,et al.  Stable Encoding of Large Finite-State Automata in Recurrent Neural Networks with Sigmoid Discriminants , 1996, Neural Computation.

[13]  Srimat T. Chakradhar,et al.  First-order versus second-order single-layer recurrent neural networks , 1994, IEEE Trans. Neural Networks.

[14]  Jack Copeland,et al.  Intelligent machinery , 2017, The Turing Guide.

[15]  J. Pollack The Induction of Dynamical Recognizers , 1996, Machine Learning.

[16]  Stefan C. Kremer,et al.  On the computational power of Elman-style recurrent networks , 1995, IEEE Trans. Neural Networks.

[17]  Alberto Sanfeliu,et al.  An Algebraic Framework to Represent Finite State Machines in Single-Layer Recurrent Neural Networks , 1995, Neural Computation.

[18]  Ofer Feinerman,et al.  Reliable neuronal logic devices from patterned hippocampal cultures , 2008 .

[19]  Alessandro E. P. Villa,et al.  An Attractor-Based Complexity Measurement for Boolean Recurrent Neural Networks , 2014, PloS one.

[20]  Hava T. Siegelmann,et al.  RECURRENT NEURAL NETWORKS AND FINITE AUTOMATA , 1996, Comput. Intell..

[21]  Yuji Ikegaya,et al.  Synfire Chains and Cortical Songs: Temporal Modules of Cortical Activity , 2004, Science.

[22]  James L. McClelland,et al.  Finite State Automata and Simple Recurrent Networks , 1989, Neural Computation.

[23]  Javier M. Buldú,et al.  Synchronization-based computation through networks of coupled oscillators , 2015, Front. Comput. Neurosci..

[24]  Fred Wolf,et al.  Neurophysics: Logic gates come to life , 2008 .

[25]  M. Abeles Local Cortical Circuits: An Electrophysiological Study , 1982 .

[26]  Hava T. Siegelmann,et al.  Computational power of neural networks: a characterization in terms of Kolmogorov complexity , 1997, IEEE Trans. Inf. Theory.

[27]  Hava T. Siegelmann,et al.  Analog computation via neural networks , 1993, [1993] The 2nd Israel Symposium on Theory and Computing Systems.

[28]  Padhraic Smyth,et al.  Learning Finite State Machines With Self-Clustering Recurrent Networks , 1993, Neural Computation.

[29]  Hava T. Siegelmann,et al.  On the Computational Power of Neural Nets , 1995, J. Comput. Syst. Sci..

[30]  Jochen Triesch,et al.  Robust development of synfire chains from multiple plasticity mechanisms , 2014, Front. Comput. Neurosci..

[31]  Alessandro E. P. Villa,et al.  Expressive power of first-order recurrent neural networks determined by their attractor dynamics , 2016, J. Comput. Syst. Sci..

[32]  Raymond L. Watrous,et al.  Induction of Finite-State Languages Using Second-Order Recurrent Networks , 1992, Neural Computation.

[33]  Aubin Tchaptchet,et al.  Modeling neuronal activity in relation to experimental voltage-/patch-clamp recordings , 2013, Brain Research.

[34]  Hava T. Siegelmann,et al.  The Dynamic Universality of Sigmoidal Neural Networks , 1996, Inf. Comput..

[35]  Hava T. Siegelmann,et al.  Neural and Super-Turing Computing , 2003, Minds and Machines.

[36]  Dong Chen,et al.  Single-Layer Recurrent Neural Networks , 1992 .

[37]  Hoppensteadt,et al.  Synchronization of laser oscillators, associative memory, and optical neurocomputing , 2000, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[38]  Jérémie Cabessa,et al.  Emulation of finite state automata with networks of synfire rings , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[39]  Jérémie Cabessa,et al.  Neural Computation with Spiking Neural Networks Composed of Synfire Rings , 2017, ICANN.

[40]  Hava T. Siegelmann,et al.  The Computational Power of Interactive Recurrent Neural Networks , 2012, Neural Computation.

[41]  S C Kleene,et al.  Representation of Events in Nerve Nets and Finite Automata , 1951 .

[42]  Stefano Boccaletti,et al.  Computation Emerges from Adaptive Synchronization of Networking Neurons , 2011, PloS one.

[43]  Hava T. Siegelmann,et al.  Turing Universality of Neural Nets (Revisited) , 1997, EUROCAST.

[44]  A. Hodgkin,et al.  A quantitative description of membrane current and its application to conduction and excitation in nerve , 1990 .

[45]  O. Prospero-Garcia,et al.  Reliability of Spike Timing in Neocortical Neurons , 1995 .

[46]  J Szentagothai,et al.  [Neuronal circuits of the cerebral cortex]. , 1970, Bulletin de l'Academie royale de medecine de Belgique.

[47]  José Carlos Príncipe,et al.  Logic computation using coupled neural oscillators , 2004, 2004 IEEE International Symposium on Circuits and Systems (IEEE Cat. No.04CH37512).

[48]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[49]  Jérémie Cabessa,et al.  Expressive Power of Nondeterministic Recurrent Neural Networks in Terms of their Attractor Dynamics , 2016, Int. J. Unconv. Comput..

[50]  Hava T. Siegelmann,et al.  The Super-Turing Computational Power of plastic Recurrent Neural Networks , 2014, Int. J. Neural Syst..

[51]  C. Lee Giles,et al.  Learning and Extracting Finite State Automata with Second-Order Recurrent Neural Networks , 1992, Neural Computation.

[52]  Hava T. Siegelmann,et al.  Evolving recurrent neural networks are super-Turing , 2011, The 2011 International Joint Conference on Neural Networks.