On-line training of recurrent neural networks with continuous topology adaptation

This paper presents an online procedure for training dynamic neural networks with input-output recurrences whose topology is continuously adjusted to the complexity of the target system dynamics. This is accomplished by changing the number of the elements of the network hidden layer whenever the existing topology cannot capture the dynamics presented by the new data. The training mechanism is based on the suitably altered extended Kalman filter (EKF) algorithm which is simultaneously used for the network parameter adjustment and for its state estimation. The network consists of a single hidden layer with Gaussian radial basis functions (GRBF), and a linear output layer. The choice of the GRBF is induced by the requirements of the online learning. The latter implies the network architecture which permits only local influence of the new data point in order not to forget the previously learned dynamics. The continuous topology adaptation is implemented in our algorithm to avoid memory and computational problems of using a regular grid of GRBF'S which covers the network input space. Furthermore, we show that the resulting parameter increase can be handled "smoothly" without interfering with the already acquired information. If the target system dynamics are changing over time, we show that a suitable forgetting factor can be used to "unlearn" the no longer-relevant dynamics. The quality of the recurrent network training algorithm is demonstrated on the identification of nonlinear dynamic systems.

[1]  Panos J. Antsaklis,et al.  An introduction to intelligent and autonomous control , 1993 .

[2]  Jay A. Farrell,et al.  A Computationally Efficient Algorithm for Training Recurrent Connectionist Networks , 1992, 1992 American Control Conference.

[3]  Lee A. Feldkamp,et al.  Decoupled extended Kalman filter training of feedforward layered networks , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[4]  Ronald J. Williams,et al.  A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.

[5]  Walter Baker,et al.  Learning control systems , 1993 .

[6]  Arthur Gelb,et al.  Applied Optimal Estimation , 1974 .

[7]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[8]  R. Palmer,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[9]  Hong-Te Su,et al.  Identification of Chemical Processes using Recurrent Networks , 1991, 1991 American Control Conference.

[10]  Sheng Chen,et al.  Reconstruction of binary signals using an adaptive radial-basis-function equalizer , 1991, Signal Process..

[11]  Ronald J. Williams,et al.  Training recurrent networks using the extended Kalman filter , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.

[12]  Jay A. Farrell,et al.  Connectionist learning control systems: submarine depth control , 1990, 29th IEEE Conference on Decision and Control.

[13]  Lennart Ljung,et al.  System Identification: Theory for the User , 1987 .

[14]  James W. Watterson An optimum multilayer perceptron neural receiver for signal detection , 1990, IEEE Trans. Neural Networks.