A unified framework for gradient algorithms used for filter adaptation and neural network training

In this paper we present in a unified framework the gradient algorithms employed in the adaptation of linear time filters (TF) and the supervised training of (non-linear) neural networks (NN). the optimality criteria used to optimize the parameters H of the filter or network are the least squares (LS) and least mean squares (LMS) in both contexts. They respectively minimize the total or the mean squares of the error e(k) between an (output) reference sequence d(k) and the actual system output y(k) corresponding to the input X(k). Minimization is performed iteratively by a gradient algorithm. the index k in (TF) is time and it runs indefinitely. Thus iterations start as soon as reception of X(k) begins. the recursive algorithm for the adaptation H(k – 1) H(k) of the parameters is implemented each time a new input X(k) is observed. When training a (NN) with a finite number of examples, the index k denotes the example and it is upper-bounded. Iterative (block) algorithms wait until all K examples are received to begin the network updating. However, K being frequently very large, recursive algorithms are also often preferred in (NN) training, but they raise the question of ordering the examples X(k). Except in the specific case of a transversal filter, there is no general recursive technique for optimizing the LS criterion. However, X(k) is normally a random stationary sequence; thus LS and LMS are equivalent when k becomes large. Moreover, the LMS criterion can always be minimized recursively with the help of the stochastic LMS gradient algorithm, which has low computational complexity. In (TF), X(k) is a sliding window of (time) samples, whereas in the supervised training of (NN) with arbitrarily ordered examples, X(k – 1) and X(k) have nothing to do with each other. When this (major) difference is rubbed out by plugging a time signal at the network input, the recursive algorithms recently developed for (NN) training become similar to those of adaptive filtering. In this context the present paper displays the similarities between adaptive cascaded linear filters and trained multilayer networks. It is also shown that there is a close similarity between adaptive recursive filters and neural networks including feedback loops. The classical filtering approach is to evaluate the gradient by ‘forward propagation’, whereas the most popular (NN) training method uses a gradient backward propagation method. We show that when a linear (TF) problem is implemented by an (NN), the two approaches are equivalent. However, the backward method can be used for more general (non-linear) filtering problems. Conversely, new insights can be drawn in the (NN) context by the use of a gradient forward computation. The advantage of the (NN) framework, and in particular of the gradient backward propagation approach, is evidently to have a much larger spectrum of applications than (TF), since (i) the inputs are arbitrary and (ii) the (NN) can perform non-linear (TF).

[1]  Maurice Bellanger,et al.  Adaptive digital filters and signal analysis , 1987 .

[2]  Bernard Widrow,et al.  Adaptive switching circuits , 1988 .

[3]  J. Shynk Adaptive IIR filtering , 1989, IEEE ASSP Magazine.

[4]  S. Stearns Error surfaces of recursive adaptive filters , 1981 .

[5]  S. T. Alexander,et al.  Adaptive Signal Processing: Theory and Applications , 1986 .

[6]  J. J. Shynk,et al.  Steady-state analysis of a single-layer perceptron based on a system identification model with bias terms , 1991 .

[7]  J. Cadzow,et al.  Signal processing via least squares error modeling , 1990, IEEE ASSP Magazine.

[8]  Charles R. Johnson,et al.  A convergence proof for a hyperstable adaptive recursive filter (Corresp.) , 1979, IEEE Trans. Inf. Theory.

[9]  S. Marcos,et al.  Joint adaptive echo cancellation and channel equalization for data transmission , 1990 .

[10]  David G. Messerschmitt,et al.  Adaptive Filters: Structures, Algorithms and Applications , 1984 .

[11]  C. Cowan,et al.  Adaptive Filters and Equalisers , 1988 .

[12]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[13]  Y. L. Cun,et al.  Modèles connexionnistes de l'apprentissage , 1987 .

[14]  C. Johnson,et al.  Theory and design of adaptive filters , 1987 .

[15]  Philip C. Treleaven,et al.  Paper: Neurocomputers , 1989 .

[16]  E. Eweda,et al.  Second-order convergence analysis of stochastic adaptive linear filtering , 1983 .

[17]  Richard P. Lippmann,et al.  An introduction to computing with neural nets , 1987 .

[18]  John J. Shynk,et al.  Performance surfaces of a single-layer perceptron , 1990, IEEE Trans. Neural Networks.

[19]  O. Macchi,et al.  When is DPCM a stable system? , 1990, International Conference on Acoustics, Speech, and Signal Processing.

[20]  S. Thomas Alexander,et al.  Adaptive Signal Processing , 1986, Texts and Monographs in Computer Science.

[21]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[22]  C. Richard Johnson,et al.  On adaptive IIR filters and parallel adaptive identifiers with adaptive error filtering , 1981, ICASSP.

[23]  Luís B. Almeida,et al.  A learning rule for asynchronous perceptrons with feedback in a combinatorial environment , 1990 .

[24]  Pineda,et al.  Generalization of back-propagation to recurrent neural networks. , 1987, Physical review letters.

[25]  Ronald J. Williams,et al.  A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.

[26]  Y. D. Landau,et al.  Adaptive control: The model reference approach , 1979, IEEE Transactions on Systems, Man, and Cybernetics.

[27]  Meriem Jaïdane,et al.  Stability of adaptive recursive filters , 1988, ICASSP-88., International Conference on Acoustics, Speech, and Signal Processing.

[28]  James S. J. Lee,et al.  A new LMS-based algorithm for rapid adaptive classification in dynamic environments: theory and preliminary results , 1988, IEEE 1988 International Conference on Neural Networks.

[29]  P. Feintuch An adaptive recursive LMS filter , 1976, Proceedings of the IEEE.

[30]  L. Personnaz,et al.  Neural network training schemes for non-linear adaptive filtering and modelling , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.