MLPs (Mono-Layer Polynomials and Multi-Layer Perceptrons) for Nonlinear Modeling

This paper presents a model selection procedure which stresses the importance of the classic polynomial models as tools for evaluating the complexity of a given modeling problem, and for removing non-significant input variables. If the complexity of the problem makes a neural network necessary, the selection among neural candidates can be performed in two phases. In an additive phase, the most important one, candidate neural networks with an increasing number of hidden neurons are trained. The addition of hidden neurons is stopped when the effect of the round-off errors becomes significant, so that, for instance, confidence intervals cannot be accurately estimated. This phase leads to a set of approved candidate networks. In a subsequent subtractive phase, a selection among approved networks is performed using statistical Fisher tests. The series of tests starts from a possibly too large unbiased network (the full network), and ends with the smallest unbiased network whose input variables and hidden neurons all have a significant contribution to the regression estimate. This method was successfully tested against the real-world regression problems proposed at the NIPS2000 Unlabeled Data Supervised Learning Competition; two of them are included here as illustrative examples.

[1]  George A. F. Seber,et al.  Linear regression analysis , 1977 .

[2]  Sheng Chen,et al.  Prediction-error estimation algorithm for non-linear output-affine systems , 1988 .

[3]  Gene H. Golub,et al.  Matrix computations , 1983 .

[4]  Hervé Stoppiglia Méthodes statistiques de sélection de modèles neuronaux ; applications financières et bancaires , 1997 .

[5]  Léon Personnaz,et al.  Construction of confidence intervals for neural networks based on least squares estimation , 2000, Neural Networks.

[6]  James T. Kwok,et al.  Objective functions for training new hidden units in constructive neural networks , 1997, IEEE Trans. Neural Networks.

[7]  I. J. Leontaritis,et al.  Model selection and validation methods for non-linear systems , 1987 .

[8]  Eric R. Ziegel,et al.  The Elements of Statistical Learning , 2003, Technometrics.

[9]  Sheng Chen,et al.  Orthogonal least squares methods and their application to non-linear system identification , 1989 .

[10]  Léon Personnaz,et al.  Jacobian Conditioning Analysis for Model Validation , 2004, Neural Computation.

[11]  John Moody,et al.  Prediction Risk and Architecture Selection for Neural Networks , 1994 .

[12]  Pascal Neveu,et al.  Bayesian nonlinear model selection and neural networks: a conjugate prior approach , 2000, IEEE Trans. Neural Networks Learn. Syst..

[13]  Gregory J. Wolff,et al.  Optimal Brain Surgeon: Extensions and performance comparisons , 1993, NIPS 1993.

[14]  N. Draper,et al.  Applied Regression Analysis. , 1967 .

[15]  Léon Personnaz,et al.  Neural-network construction and selection in nonlinear modeling , 2003, IEEE Trans. Neural Networks.

[16]  James T. Kwok,et al.  Constructive algorithms for structure learning in feedforward neural networks for regression problems , 1997, IEEE Trans. Neural Networks.

[17]  Léon Personnaz,et al.  On Cross Validation for Model Selection , 1999, Neural Computation.

[18]  Yann LeCun,et al.  Optimal Brain Damage , 1989, NIPS.

[19]  Ulrich Anders,et al.  Model selection in neural networks , 1999, Neural Networks.

[20]  Léon Personnaz,et al.  A statistical procedure for determining the optimal number of hidden neurons of a neural model , 2000 .

[21]  R. Kass Nonlinear Regression Analysis and its Applications , 1990 .

[22]  Léon Personnaz,et al.  CONSTRUCTION OF CONFIDENCE INTERVALS IN NEURAL MODELING USING A LINEAR TAYLOR EXPANSION , 1998 .