An efficient training algorithm for dynamic synapse neural networks using trust region methods

We formulate the dynamic synapse neural network from the averaged activity of local population of neurons perspective. We have applied the trust region nonlinear optimization approach to train the network and show the new learning method effectiveness in comparison to the genetic algorithms by optimizing large-scale networks.

[1]  Wolfgang Maass,et al.  Spiking Neurons , 1998, NC.

[2]  Theodore W. Berger,et al.  A new dynamic synapse neural network for speech recognition , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).

[3]  E. Izhikevich,et al.  Weakly connected neural networks , 1997 .

[4]  Thomas F. Coleman,et al.  An Interior Trust Region Approach for Nonlinear Minimization Subject to Bounds , 1993, SIAM J. Optim..

[5]  Wulfram Gerstner,et al.  Spiking neurons , 1999 .

[6]  Bing J. Sheu,et al.  Brain-implantable biomimetic electronics as the next era in neural prosthetics , 2001, Proc. IEEE.

[7]  T. Coleman,et al.  On the Convergence of Reflective Newton Methods for Large-scale Nonlinear Minimization Subject to Bounds , 1992 .

[8]  S. Mallat A wavelet tour of signal processing , 1998 .

[9]  T.W. Berger,et al.  The Gauss-Newton learning method for a generalized dynamic synapse neural network , 2002, Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290).

[10]  J S Liaw,et al.  Dynamic synapse: A new concept of neural representation and computation , 1996, Hippocampus.

[11]  William Bialek,et al.  Spikes: Exploring the Neural Code , 1996 .

[12]  Christopher J. Bishop,et al.  Pulsed Neural Networks , 1998 .

[13]  Thomas F. Coleman,et al.  On the convergence of interior-reflective Newton methods for nonlinear minimization subject to bounds , 1994, Math. Program..

[14]  W. Pitts,et al.  A Logical Calculus of the Ideas Immanent in Nervous Activity (1943) , 2021, Ideas That Created the Future.