Open-loop training of recurrent neural networks for nonlinear dynamical system identification

We develop a training approach for a class of recurrent neural networks which are categorized by layered links from input neurons to output neurons and time-lagged feedback links from output neurons to input neurons. This particular neural network structure can be considered as a special case of time-lagged recurrent networks. The present approach treats the recurrent neural network as a multilayer feedforward neural network during training by opening up the feedback links. We also treat the nonlinear system to be identified as a nonlinear function with no dynamics during data collection. Such a process for training data collection allows the use of random system states and random control inputs to ensure good representation in data collection and less dependence on the initial states. The training process of the neural networks can be simplified since the gradient calculation is much less involved in feedforward neural networks. It is argued that the neural network structure considered herein is appropriate for performing nonlinear dynamical system identification.

[1]  Jacob Barhen,et al.  Learning a trajectory using adjoint functions and teacher forcing , 1992, Neural Networks.

[2]  Jun Wang,et al.  On-line learning of dynamical systems in the presence of model mismatch and disturbances , 2000, IEEE Trans. Neural Networks Learn. Syst..

[3]  Kumpati S. Narendra,et al.  On a new approach to the design of tracking controllers for nonlinear dynamical systems , 1999, Proceedings of the 1999 American Control Conference (Cat. No. 99CH36251).

[4]  Barak A. Pearlmutter Gradient calculations for dynamic recurrent neural networks: a survey , 1995, IEEE Trans. Neural Networks.

[5]  Lee A. Feldkamp,et al.  Dynamic neural network methods applied to on-vehicle idle speed control , 1996, Proc. IEEE.

[6]  Ronald J. Williams,et al.  A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.

[7]  Paul J. Werbos,et al.  Approximate dynamic programming for real-time control and neural modeling , 1992 .

[8]  Lee A. Feldkamp,et al.  Neurocontrol of nonlinear dynamical systems with Kalman filter trained recurrent networks , 1994, IEEE Trans. Neural Networks.

[9]  G. V. Puskorius,et al.  Truncated backpropagation through time and Kalman filter training for neurocontrol , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).

[10]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[11]  Amir F. Atiya,et al.  New results on recurrent network training: unifying the algorithms and accelerating convergence , 2000, IEEE Trans. Neural Networks Learn. Syst..

[12]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[13]  Jürgen Schmidhuber,et al.  A Fixed Size Storage O(n3) Time Complexity Learning Algorithm for Fully Recurrent Continually Running Networks , 1992, Neural Computation.

[14]  Kumpati S. Narendra,et al.  Identification and control of dynamical systems using neural networks , 1990, IEEE Trans. Neural Networks.

[15]  Jing Peng,et al.  An Efficient Gradient-Based Algorithm for On-Line Training of Recurrent Network Trajectories , 1990, Neural Computation.

[16]  Lee A. Feldkamp,et al.  Adaptive Behavior from Fixed Weight Networks , 1997, Inf. Sci..