A New Jacobian Matrix for Optimal Learning of Single-Layer Neural Networks

This paper investigates the learning of a wide class of single-hidden-layer feedforward neural networks (SLFNs) with two sets of adjustable parameters, i.e., the nonlinear parameters in the hidden nodes and the linear output weights. The main objective is to both speed up the convergence of second-order learning algorithms such as Levenberg-Marquardt (LM), as well as to improve the network performance. This is achieved here by reducing the dimension of the solution space and by introducing a new Jacobian matrix. Unlike conventional supervised learning methods which optimize these two sets of parameters simultaneously, the linear output weights are first converted into dependent parameters, thereby removing the need for their explicit computation. Consequently, the neural network (NN) learning is performed over a solution space of reduced dimension. A new Jacobian matrix is then proposed for use with the popular second-order learning methods in order to achieve a more accurate approximation of the cost function. The efficacy of the proposed method is shown through an analysis of the computational complexity and by presenting simulation results from four different examples.

[1]  Kenneth Levenberg A METHOD FOR THE SOLUTION OF CERTAIN NON – LINEAR PROBLEMS IN LEAST SQUARES , 1944 .

[2]  Marco Gori,et al.  On the problem of local minima in recurrent neural networks , 1994, IEEE Trans. Neural Networks.

[3]  Jorge J. Moré,et al.  The Levenberg-Marquardt algo-rithm: Implementation and theory , 1977 .

[4]  George W. Irwin,et al.  A hybrid linear/nonlinear training algorithm for feedforward neural networks , 1998, IEEE Trans. Neural Networks.

[5]  F. Lewis,et al.  Guest editorial: Neural network feedback control with guaranteed stability , 1998 .

[6]  Henry Leung,et al.  Prediction of noisy chaotic time series using an optimal radial basis function neural network , 2001, IEEE Trans. Neural Networks.

[7]  Shiro Usui,et al.  Mutation-based genetic neural network , 2005, IEEE Transactions on Neural Networks.

[8]  Benjamin W. Wah,et al.  Global Optimization for Neural Network Training , 1996, Computer.

[9]  Michael J. Korenberg,et al.  Iterative fast orthogonal search algorithm for MDL-based training of generalized single-layer networks , 2000, Neural Networks.

[10]  David McLean,et al.  On Global–Local Artificial Neural Networks for Function Approximation , 2006, IEEE Transactions on Neural Networks.

[11]  Nicolaos B. Karayiannis,et al.  Reformulated radial basis neural networks trained by gradient descent , 1999, IEEE Trans. Neural Networks.

[12]  Roberto Battiti,et al.  First- and Second-Order Methods for Learning: Between Steepest Descent and Newton's Method , 1992, Neural Computation.

[13]  Kok Kiong Tan,et al.  Nonlinear adaptive control of interconnected systems using neural networks , 2006, IEEE Transactions on Neural Networks.

[14]  George W. Irwin,et al.  A fast nonlinear model identification method , 2005, IEEE Transactions on Automatic Control.

[15]  D. Marquardt An Algorithm for Least-Squares Estimation of Nonlinear Parameters , 1963 .

[16]  Tung-Kuan Liu,et al.  Tuning the structure and parameters of a neural network by using hybrid Taguchi-genetic algorithm , 2006, IEEE Trans. Neural Networks.

[17]  Russell Reed,et al.  Pruning algorithms-a survey , 1993, IEEE Trans. Neural Networks.

[18]  Mohammad Bagher Menhaj,et al.  Training feedforward networks with the Marquardt algorithm , 1994, IEEE Trans. Neural Networks.

[19]  Chee Kheong Siew,et al.  Can threshold networks be trained directly? , 2006, IEEE Transactions on Circuits and Systems II: Express Briefs.

[20]  Kang Li,et al.  A two-stage algorithm for identification of nonlinear dynamic systems , 2006, Autom..

[21]  Bor-Shing Lin,et al.  Higher-Order-Statistics-Based Radial Basis Function Networks for Signal Enhancement , 2007, IEEE Transactions on Neural Networks.

[22]  Jacek M. Zurada,et al.  An energy function-based design method for discrete hopfield associative memory with attractive fixed points , 2005, IEEE Transactions on Neural Networks.

[23]  Yukihiro Toyoda,et al.  A parameter optimization method for radial basis function type models , 2003, IEEE Trans. Neural Networks.

[24]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[25]  Stefanos Kollias,et al.  An adaptive least squares algorithm for the efficient training of artificial neural networks , 1989 .

[26]  Marios M. Polycarpou,et al.  Neural network based fault detection in robotic manipulators , 1998, IEEE Trans. Robotics Autom..

[27]  Simon X. Yang,et al.  A Neural Network Approach to Dynamic Task Assignment of Multirobots , 2006, IEEE Transactions on Neural Networks.

[28]  De-Shuang Huang,et al.  A Hybrid Forward Algorithm for RBF Neural Network Construction , 2006, IEEE Transactions on Neural Networks.

[29]  Shubao Liu,et al.  A Simplified Dual Neural Network for Quadratic Programming With Its KWTA Application , 2006, IEEE Transactions on Neural Networks.

[30]  Dipti Srinivasan,et al.  Neural Networks for Continuous Online Learning and Control , 2006, IEEE Transactions on Neural Networks.

[31]  Michel Verleysen,et al.  Width optimization of the Gaussian kernels in Radial Basis Function Networks , 2002, ESANN.

[32]  D. Gorinevsky An approach to parametric nonlinear least square optimization and application to task-level learning control , 1997, IEEE Trans. Autom. Control..

[33]  Sheng Chen,et al.  Orthogonal least squares methods and their application to non-linear system identification , 1989 .

[34]  Narasimhan Sundararajan,et al.  A generalized growing and pruning RBF (GGAP-RBF) neural network for function approximation , 2005, IEEE Transactions on Neural Networks.

[35]  Zekeriya Uykan,et al.  Analysis of input-output clustering for determining centers of RBFN , 2000, IEEE Trans. Neural Networks Learn. Syst..

[36]  Roman Neruda,et al.  Learning methods for radial basis function networks , 2005, Future Gener. Comput. Syst..

[37]  Tao Zhang,et al.  Stable Adaptive Neural Network Control , 2001, The Springer International Series on Asian Studies in Computer and Information Science.

[38]  Jooyoung Park,et al.  Universal Approximation Using Radial-Basis-Function Networks , 1991, Neural Computation.

[39]  Kevin Warwick,et al.  Mean-tracking clustering algorithm for radial basis function centre selection , 1997 .

[40]  Robert M. Sanner,et al.  Gaussian Networks for Direct Adaptive Control , 1991, 1991 American Control Conference.

[41]  George D. Magoulas,et al.  Deterministic nonmonotone strategies for effective training of multilayer perceptrons , 2002, IEEE Trans. Neural Networks.

[42]  Griff L. Bilbro Fast stochastic global optimization , 1993, Optics & Photonics.

[43]  Henry Leung,et al.  Blind equalization using a predictive radial basis function neural network , 2005, IEEE Trans. Neural Networks.

[44]  Stavros J. Perantonis,et al.  Two highly efficient second-order algorithms for training feedforward networks , 2002, IEEE Trans. Neural Networks.