A tree-structured adaptive network for function approximation in high-dimensional spaces

Nonlinear function approximation is often solved by finding a set of coefficients for a finite number of fixed nonlinear basis functions. However, if the input data are drawn from a high-dimensional space, the number of required basis functions grows exponentially with dimension, leading many to suggest the use of adaptive nonlinear basis functions whose parameters can be determined by iterative methods. The author proposes a technique based on the idea that for most of the data, only a few dimensions of the input may be necessary to compute the desired output function. Additional input dimensions are incorporated only where needed. The learning procedure grows a tree whose structure depends upon the input data and the function to be approximated. This technique has a fast learning algorithm with no local minima once the network shape is fixed, and it can be used to reduce the number of required measurements in situations where there is a cost associated with sensing. Three examples are given: controlling the dynamics of a simulated planar two-joint robot arm, predicting the dynamics of the chaotic Mackey-Glass equation, and predicting pixel values in real images from pixel values above and to the left.

[1]  Dennis Gabor,et al.  A universal nonlinear filter, predictor and simulator which optimizes itself by a learning process , 1961 .

[2]  A. A. Mullin,et al.  Principles of neurodynamics , 1962 .

[3]  J. Morgan,et al.  Problems in the Analysis of Survey Data, and a Proposal , 1963 .

[4]  A. G. Ivakhnenko,et al.  Polynomial Theory of Complex Systems , 1971, IEEE Trans. Syst. Man Cybern..

[5]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[6]  A. Isidori,et al.  Realization and Structure Theory of Bilinear Dynamical Systems , 1974 .

[7]  Jon Louis Bentley,et al.  Multidimensional binary search trees used for associative searching , 1975, CACM.

[8]  B. Widrow,et al.  Stationary and nonstationary learning characteristics of the LMS adaptive filter , 1976, Proceedings of the IEEE.

[9]  Saburo Ikeda,et al.  Sequential GMDH Algorithm and Its Application to River Flow Prediction , 1976 .

[10]  Lennart Ljung,et al.  Analysis of recursive stochastic algorithms , 1977 .

[11]  L. Glass,et al.  Oscillation and chaos in physiological control systems. , 1977, Science.

[12]  Abraham Lempel,et al.  Compression of individual sequences via variable-rate coding , 1978, IEEE Trans. Inf. Theory.

[13]  C. J. Stone,et al.  Additive Regression and Other Nonparametric Models , 1985 .

[14]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[15]  Robert M. Farber,et al.  How Neural Nets Work , 1987, NIPS.

[16]  Guo-Zheng Sun,et al.  A Novel Net that Learns Sequential Decision Process , 1987, NIPS.

[17]  A. Lapedes,et al.  Nonlinear Signal Processing Using Neural Networks , 1987 .

[18]  M. J. D. Powell,et al.  Radial basis functions for multivariable interpolation: a review , 1987 .

[19]  Bernard Widrow,et al.  Adaptive switching circuits , 1988 .

[20]  Michael F. Shlesinger,et al.  Dynamic patterns in complex systems , 1988 .

[21]  W. Cleveland,et al.  Regression by local fitting: Methods, properties, and computational algorithms , 1988 .

[22]  David S. Broomhead,et al.  Multivariable Functional Interpolation and Adaptive Networks , 1988, Complex Syst..

[23]  W. Cleveland,et al.  Locally Weighted Regression: An Approach to Regression Analysis by Local Fitting , 1988 .

[24]  John E. Moody,et al.  Fast Learning in Multi-Resolution Hierarchies , 1988, NIPS.

[25]  D. Broomhead,et al.  Radial Basis Functions, Multi-Variable Functional Interpolation and Adaptive Networks , 1988 .

[26]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[27]  Gérard Dreyfus,et al.  Single-layer learning revisited: a stepwise procedure for building and training a neural network , 1989, NATO Neurocomputing.

[28]  M. F. Tenorio,et al.  Self-Organizing Neural Network for Optimum Supervised Learning , 1989 .

[29]  Jae S. Lim,et al.  Two-Dimensional Signal and Image Processing , 1989 .

[30]  T Poggio,et al.  Regularization Algorithms for Learning That Are Equivalent to Multilayer Networks , 1990, Science.

[31]  J. Friedman Multivariate adaptive regression splines , 1990 .

[32]  T. Sanger Basis-Function Trees for Approximation in High-Dimensional Spaces , 1991 .

[33]  J. Freidman,et al.  Multivariate adaptive regression splines , 1991 .

[34]  Terence D. Sanger,et al.  A Tree-Structured Algorithm for Reducing Computation in Networks with Separable Basis Functions , 1991, Neural Computation.

[35]  Michael I. Jordan,et al.  Task Decomposition Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks , 1990, Cogn. Sci..