Model structure determination in neural network models

Feedforward neural networks (FNN) have been used intensively for the identification and control of chemical engineering processes. However, there is no efficient model structure determination methodology for a particular mapping application. This has resulted in a tendency to use networks that are much larger than required. In this paper a new procedure for model structure determination in feedforward neural networks is proposed. This procedure is based on network pruning using the orthogonal least-squares technique to determine insignificant or redundant synaptic weights, biases, hidden nodes and network inputs. The advantages of this approach are discussed and illustrated using simulated and experimental data. The results show that the orthogonal least-squares technique is quite efficient in determining the significant elements on the neural network models. The results also show the importance of pruning procedures to identify parsimonious FNN models.

[1]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1992, Math. Control. Signals Syst..

[2]  Yann LeCun,et al.  Optimal Brain Damage , 1989, NIPS.

[3]  Sheng Chen,et al.  Identification of MIMO non-linear systems using a forward-regression orthogonal estimator , 1989 .

[4]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..

[5]  Milan Kubíček,et al.  Book-Review - Computational Methods in Bifurcation Theory and Dissipative Structures , 1983 .

[6]  S. Billings,et al.  Orthogonal parameter estimation algorithm for non-linear stochastic systems , 1988 .

[7]  Dale E. Seborg,et al.  Determination of model order for NARX models directly from input-output data , 1998 .

[8]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[9]  P. Werbos,et al.  Long-term predictions of chemical processes using recurrent neural networks: a parallel training approach , 1992 .

[10]  Martin Pottman,et al.  Identification of non-linear processes using reciprocal multiquadric functions , 1992 .

[11]  Joseph A. Wolkan,et al.  Introduction to probability and statistics , 1994 .

[12]  Michael Nikolaou,et al.  Dynamic process modeling with recurrent neural networks , 1993 .

[13]  T. Gustafsson,et al.  Dynamic modeling and reaction invariant control of pH , 1983 .

[14]  Gregory J. Wolff,et al.  Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.

[15]  Michael C. Mozer,et al.  Skeletonization: A Technique for Trimming the Fat from a Network via Relevance Assessment , 1988, NIPS.

[16]  A. Hindmarsh LSODE and LSODI, two new initial value ordinary differential equation solvers , 1980, SGNM.

[17]  J. S. Milton,et al.  Introduction to Probability and Statistics: Principles and Applications for Engineering and the Comp , 1995 .

[18]  Dale E. Seborg,et al.  Modelling and Self-Tuning Control of a Multivariable pH Neutralization Process Part I: Modelling and Multiloop Control , 1989, 1989 American Control Conference.

[19]  Masafumi Hagiwara Novel backpropagation algorithm for reduction of hidden units and acceleration of convergence using artificial selection , 1990, 1990 IJCNN International Joint Conference on Neural Networks.

[20]  N. V. Bhat,et al.  determining model structure for neural models by network stripping , 1992 .