Efficient learning procedures for optimal interpolative nets

The Optimal Interpolative (OI) net, derived earlier by one of the authors based on a generalized Fock (GF) space formulation, provides a closed form solution to the synaptic weights in a two-layer feedforward neural network. In this paper, a recursive learning algorithm called RLS-OI is introduced, which extends the original OI solution to a recursive least squares solution. This is an enhanced version of a learning procedure we derived recently also based on the OI solution, which brings in new features such as a procedure for the selection of a good training sequence and a scaling factor in the nonlinear activation function to the hidden neurons. A more general approach is used in its derivation which makes it possible to apply the same technique with different forms of activation functions or basis functions. The proposed algorithm selects a small but meaningful (consistent) subset of exemplars known as prototypes and recursively updates the synaptic weights of the neural network as each new prototype is found. RLS-OI is a noniterative algorithm and is guaranteed to produce a solution in a number of recursions no larger than the number of exemplars in the training set. It also requires no prespecification of the architecture of the neural network to be learned but evolves a net from scratch to the extent sufficient to solve a given problem. The method for picking prototypes and the recursive relations needed to implement the algorithm are described in detail. Simulation results are provided to illustrate our approach.

[1]  C. G. Hilborn,et al.  The Condensed Nearest Neighbor Rule , 1967 .

[2]  Alan J. Miller Subset Selection in Regression , 1992 .

[3]  R. D. Figueiredo A generalized Fock space framework for nonlinear system and signal analysis , 1983 .

[4]  Gene H. Golub,et al.  Matrix computations , 1983 .

[5]  R. J. P. de Figueiredo,et al.  An optimal matching-score net for pattern classification , 1990, IJCNN.

[6]  Alan J. Miller,et al.  Subset Selection in Regression , 1991 .

[7]  Rui J. P. de Figueiredo,et al.  A best approximation framework and implementation for simulation of large-scale nonlinear systems , 1980 .

[8]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[9]  G. Krishna,et al.  The condensed nearest neighbor rule using the concept of mutual nearest neighborhood (Corresp.) , 1979, IEEE Trans. Inf. Theory.

[10]  S. T. Alexander,et al.  Adaptive Signal Processing: Theory and Applications , 1986 .

[11]  K. Lang,et al.  Learning to tell two spirals apart , 1988 .

[12]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[13]  J. D. Powell,et al.  Radial basis function approximations to polynomials , 1989 .

[14]  I. Tomek,et al.  Two Modifications of CNN , 1976 .

[15]  Rui J. P. de Figueiredo,et al.  An evolution-oriented learning algorithm for the optimal interpolative net , 1992, IEEE Trans. Neural Networks.

[16]  James L. McClelland,et al.  James L. McClelland, David Rumelhart and the PDP Research Group, Parallel distributed processing: explorations in the microstructure of cognition . Vol. 1. Foundations . Vol. 2. Psychological and biological models . Cambridge MA: M.I.T. Press, 1987. , 1989, Journal of Child Language.

[17]  Chin-Liang Chang,et al.  Finding Prototypes For Nearest Neighbor Classifiers , 1974, IEEE Transactions on Computers.

[18]  R.J.P. deFigueirado,et al.  A new nonlinear functional analytic framework for modeling artificial neural networks , 1990, IEEE International Symposium on Circuits and Systems.

[19]  Peter E. Hart,et al.  The condensed nearest neighbor rule (Corresp.) , 1968, IEEE Trans. Inf. Theory.