Enhancing MLP networks using a distributed data representation

Multilayer perceptron (MLP) networks trained using backpropagation can be slow to converge in many instances. The primary reason for slow learning is the global nature of backpropagation. Another reason is the fact that a neuron in an MLP network functions as a hyperplane separator and is therefore inefficient when applied to classification problems in which decision boundaries are nonlinear. This paper presents a data representational approach that addresses these problems while operating within the framework of the familiar backpropagation model. We examine the use of receptors with overlapping receptive fields as a preprocessing technique for encoding inputs to MLP networks. The proposed data representation scheme, termed ensemble encoding, is shown to promote local learning and to provide enhanced nonlinear separability. Simulation results for well known problems in classification and time-series prediction indicate that the use of ensemble encoding can significantly reduce the time required to train MLP networks. Since the choice of representation for input data is independent of the learning algorithm and the functional form employed in the MLP model, nonlinear preprocessing of network inputs may be an attractive alternative for many MLP network applications.

[1]  James A. Anderson Data representation in neural networks , 1990 .

[2]  James S. Albus,et al.  New Approach to Manipulator Control: The Cerebellar Model Articulation Controller (CMAC)1 , 1975 .

[3]  Paul J. Werbos,et al.  Links Between Artificial Neural Networks (ANN) and Statistical Pattern Recognition , 1991 .

[4]  L. Glass,et al.  Oscillation and chaos in physiological control systems. , 1977, Science.

[5]  A. Gallant,et al.  Finding Chaos in Noisy Systems , 1992 .

[6]  J. D. Farmer,et al.  Chaotic attractors of an infinite-dimensional dynamical system , 1982 .

[7]  John Paul Brady,et al.  Neural bases of behavior. , 1963 .

[8]  Kumpati S. Narendra,et al.  Adaptive control using neural networks , 1990 .

[9]  Martin Casdagli,et al.  Nonlinear prediction of chaotic time series , 1989 .

[10]  Stephen H. Lane,et al.  Multi-Layer Perceptrons with B-Spline Receptive Field Functions , 1990, NIPS.

[11]  Daniel M. Wolpert,et al.  Detecting chaos with neural networks , 1990, Proceedings of the Royal Society of London. Series B: Biological Sciences.

[12]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..

[13]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[14]  Farmer,et al.  Predicting chaotic time series. , 1987, Physical review letters.

[15]  Terrence J. Sejnowski,et al.  Analysis of hidden units in a layered network trained to classify sonar targets , 1988, Neural Networks.

[16]  Vladimir Cherkassky,et al.  Data representation for diagnostic neural networks , 1992, IEEE Expert.

[17]  D.A. Handelman,et al.  Theory and development of higher-order CMAC neural networks , 1992, IEEE Control Systems.

[18]  T Poggio,et al.  Regularization Algorithms for Learning That Are Equivalent to Multilayer Networks , 1990, Science.

[19]  Peter J B Hancock,et al.  Coding Strategies for Genetic Algorithms and Neural Nets , 1993 .