Cepstral coefficients, covariance lags, and pole-zero models for finite data strings

One of the most widely used methods of spectral estimation in signal and speech processing is linear predictive coding (LPC), LPC has some attractive features, which account for its popularity, including the properties that the resulting modeling filter (i) matches a finite window of n+1 covariance lags, (ii) is rational of degree at most n, and (iii) has stable zeros and poles. The only limiting factor of this methodology is that the modeling filter is "all-pole," i.e., an autoregressive (AR) model. In this paper, we present a systematic description of all autoregressive moving-average (ARMA) models of processes that have properties (i)-(iii) in the context of cepstral analysis and homomorphic filtering. We show that each such an ARMA model determines and is completely determined by its finite windows of cepstral coefficients and covariance lags. We show that these nth-order windows form local coordinates for all ARMA models of degree n and that the pole-zero model can be determined from the windows as the unique minimum of a convex objective function. We refine this optimization method by first noting that the maximum entropy design of an LPC filter is obtained by maximizing the zeroth cepstral coefficient, subject to the constraint (i). More generally, we modify this scheme to a more well-posed optimization problem where the covariance data enter as a constraint and the linear weights of the cepstral coefficients are "positive"-in a sense that a certain pseudo-polynomial is positive-rather succinctly generalizing the maximum entropy method. This new problem is a homomorphic filter generalization of the maximum entropy method. providing a procedure for the design of any stable, minimum-phase modeling filter of degree less or equal to n that interpolates the given covariance window. We present an algorithm for realizing these filters in a lattice-ladder form, given the covariance window and the moving average part of the model, While we also show how to determine the moving average part using cepstral smoothing, one can make use of any good a priori estimate for the system zeros to initialize the algorithm. We conclude the paper with an example of this method, incorporating an example from the literature on ARMA modeling.

[1]  Genshiro Kitagawa,et al.  State Space Modeling of Time Series , 1994 .

[2]  Petre Stoica,et al.  Decentralized Control , 2018, The Control Systems Handbook.

[3]  José M. F. Moura,et al.  Parametric spectral estimation for arma processes , 1986 .

[4]  C. Byrnes,et al.  A Convex Optimization Approach to the Rational Covariance Extension Problem , 1999 .

[5]  Anders Lindquist,et al.  Experimental evidence showing that stochastic subspace identification methods may fail 1 1 This rese , 1998 .

[6]  J.M.F. Moura,et al.  LD2-ARMA identification algorithm , 1991, IEEE Trans. Signal Process..

[7]  Tryphon T. Georgiou,et al.  Realization of power spectra from partial covariance sequences , 1987, IEEE Trans. Acoust. Speech Signal Process..

[8]  Michel Minoux,et al.  Mathematical Programming , 1986 .

[9]  Paul Van Dooren,et al.  Speech modelling and the trigonometric moment problem , 1982 .

[10]  Bart De Moor,et al.  Subspace algorithms for the stochastic identification problem, , 1993, Autom..

[11]  Paul Van Dooren,et al.  On the role of the partial trigonometric moment problem in AR speech modelling , 1983, ICASSP.

[12]  Josiane Zerubia,et al.  Comparison of two ARMA estimators , 1989, International Conference on Acoustics, Speech, and Signal Processing,.

[13]  K. Åström Introduction to Stochastic Control Theory , 1970 .

[14]  Richard J. Mammone,et al.  Robust Speech Processing as an Inverse Problem , 1999 .

[15]  L. Jackson Digital filters and signal processing , 1985 .

[16]  G. Wilson Factorization of the Covariance Generating Function of a Pure Moving Average Process , 1969 .

[17]  R. Kálmán Realization of Covariance Sequences , 1982 .

[18]  John E. Markel,et al.  Linear Prediction of Speech , 1976, Communication and Cybernetics.

[19]  Georg Heinig,et al.  Fast inversion algorithms of Toeplitz-plus-Hankel matrices , 1988 .

[20]  Anders Lindquist,et al.  Geometry of the Kimura-Georgiou parametrization of modelling filters , 1989 .

[21]  H. Landau Moments in mathematics , 1987 .

[22]  J. Makhoul,et al.  Linear prediction: A tutorial review , 1975, Proceedings of the IEEE.

[23]  M.G. Bellanger,et al.  Digital processing of speech signals , 1980, Proceedings of the IEEE.

[24]  Alan V. Oppenheim,et al.  Speech analysis homomorphic prediction , 1977 .

[25]  Anders Lindquist,et al.  Canonical correlation analysis, approximate covariance extension, and identification of stationary time series , 1996, Autom..

[26]  J. Lumley,et al.  Processing of Random Signals , 1978 .

[27]  C. Byrnes,et al.  On the partial stochastic realization problem , 1997, IEEE Trans. Autom. Control..

[28]  K. Stevens,et al.  Reduction of Speech Spectra by Analysis‐by‐Synthesis Techniques , 1961 .

[29]  J. Cadzow,et al.  Spectral estimation: An overdetermined rational model equation approach , 1982, Proceedings of the IEEE.

[30]  K. Steiglitz On the simultaneous estimation of poles and zeros in speech analysis , 1977 .

[31]  J. Zerubia,et al.  A method to estimate the parameters of an ARMA model , 1987 .

[32]  Alan V. Oppenheim,et al.  Signal analysis by homomorphic prediction , 1976 .

[33]  P. Fuhrmann A Polynomial Approach to Linear Algebra , 1996 .

[34]  Stephen A. Dyer,et al.  Digital signal processing , 2018, 8th International Multitopic Conference, 2004. Proceedings of INMIC 2004..

[35]  C. Byrnes,et al.  A complete parameterization of all positive rational extensions of a covariance sequence , 1995, IEEE Trans. Autom. Control..

[36]  Bishnu S. Atal,et al.  Current methods of digital speech processing , 1990 .

[37]  Petre Stoica,et al.  Introduction to spectral analysis , 1997 .