Compressive system identification: Sequential methods and entropy bounds

In the first part of this work, a novel Kalman filtering-based method is introduced for estimating the coefficients of sparse, or more broadly, compressible autoregressive models using fewer observations than normally required. By virtue of its (unscented) Kalman filter mechanism, the derived method essentially addresses the main difficulties attributed to the underlying estimation problem. In particular, it facilitates sequential processing of observations and is shown to attain a good recovery performance, particularly under substantial deviations from ideal conditions, those which are assumed to hold true by the theory of compressive sensing. In the remaining part of this paper we derive a few information-theoretic bounds pertaining to the problem at hand. The obtained bounds establish the relation between the complexity of the autoregressive process and the attainable estimation accuracy through the use of a novel measure of complexity. This measure is used in this work as a substitute to the generally incomputable restricted isometric property.

[1]  Y. C. Pati,et al.  Orthogonal matching pursuit: recursive function approximation with applications to wavelet decomposition , 1993, Proceedings of 27th Asilomar Conference on Signals, Systems and Computers.

[2]  Sheng Chen,et al.  Orthogonal least squares methods and their application to non-linear system identification , 1989 .

[3]  Richard G. Baraniuk,et al.  Democracy in Action: Quantization, Saturation, and Compressive Sensing , 2011 .

[4]  Stéphane Mallat,et al.  Matching pursuits with time-frequency dictionaries , 1993, IEEE Trans. Signal Process..

[5]  C. Granger Investigating causal relations by econometric models and cross-spectral methods , 1969 .

[6]  Tyrone L. Vincent,et al.  Compressive System Identification of LTI and LTV ARX models , 2011, IEEE Conference on Decision and Control and European Control Conference.

[7]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[8]  Lawrence Carin,et al.  Bayesian Compressive Sensing , 2008, IEEE Transactions on Signal Processing.

[9]  Dimitri Kanevsky,et al.  A Simple Method for Sparse Signal Recovery from Noisy Observations Using Kalman Filtering , 2008 .

[10]  Joel A. Tropp,et al.  Greed is good: algorithmic results for sparse approximation , 2004, IEEE Transactions on Information Theory.

[11]  M. Rudelson,et al.  Geometric approach to error-correcting codes and reconstruction of signals , 2005, math/0502299.

[12]  A. Sokal Monte Carlo Methods in Statistical Mechanics: Foundations and New Algorithms , 1997 .

[13]  E.J. Candes,et al.  An Introduction To Compressive Sampling , 2008, IEEE Signal Processing Magazine.

[14]  George Eastman House,et al.  Sparse Bayesian Learning and the Relevan e Ve tor Ma hine , 2001 .

[15]  Dimitri Kanevsky,et al.  Unscented compressed sensing , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[16]  Justin K. Romberg,et al.  Dynamic updating for sparse time varying signals , 2009, 2009 43rd Annual Conference on Information Sciences and Systems.

[17]  Michèle Basseville,et al.  The asymptotic local approach to change detection and model validation , 1987 .

[18]  Justin K. Romberg,et al.  Sparsity penalties in dynamical system estimation , 2011, 2011 45th Annual Conference on Information Sciences and Systems.

[19]  Emmanuel J. Candès,et al.  Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.

[20]  E. Candès The restricted isometry property and its implications for compressed sensing , 2008 .

[21]  Joel A. Tropp,et al.  Signal Recovery From Random Measurements Via Orthogonal Matching Pursuit , 2007, IEEE Transactions on Information Theory.

[22]  H. Akaike A new look at the statistical model identification , 1974 .

[23]  Yonina C. Eldar,et al.  Compressed Sensing with Coherent and Redundant Dictionaries , 2010, ArXiv.

[24]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[25]  Alan Pankratz,et al.  Forecasting with univariate Box-Jenkins models : concepts and cases , 1983 .

[26]  Georgios B. Giannakis,et al.  Online Adaptive Estimation of Sparse Signals: Where RLS Meets the $\ell_1$ -Norm , 2010, IEEE Transactions on Signal Processing.

[27]  Nir Friedman,et al.  Learning Bayesian Network Structure from Massive Datasets: The "Sparse Candidate" Algorithm , 1999, UAI.

[28]  Vahid Tarokh,et al.  Adaptive algorithms for sparse system identification , 2011, Signal Process..

[29]  Pini Gurfil,et al.  Methods for Sparse Signal Recovery Using Kalman Filtering With Embedded Pseudo-Measurement Norms and Quasi-Norms , 2010, IEEE Transactions on Signal Processing.

[30]  Thomas Kailath,et al.  Linear Systems , 1980 .

[31]  Jeffrey K. Uhlmann,et al.  New extension of the Kalman filter to nonlinear systems , 1997, Defense, Security, and Sensing.

[32]  Richard A. Davis,et al.  Time Series: Theory and Methods , 2013 .

[33]  David L Donoho,et al.  Compressed sensing , 2006, IEEE Transactions on Information Theory.

[34]  Gareth M. James,et al.  DASSO: connections between the Dantzig selector and lasso , 2009 .

[35]  Stefan Haufe,et al.  Sparse Causal Discovery in Multivariate Time Series , 2008, NIPS Causality: Objectives and Assessment.

[36]  Justin K. Romberg,et al.  Estimation and dynamic updating of time-varying signals with sparse variations , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[37]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[38]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[39]  G. Giannakis,et al.  Compressed sensing of time-varying signals , 2009, 2009 16th International Conference on Digital Signal Processing.

[40]  Itzhack Y. Bar-Itzhack,et al.  Quaternion normalization in spacecraft attitude determination , 1992 .

[41]  Mike E. Davies,et al.  Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.

[42]  Terence Tao,et al.  The Dantzig selector: Statistical estimation when P is much larger than n , 2005, math/0506081.

[43]  Joseph J. LaViola,et al.  On Kalman Filtering With Nonlinear Equality Constraints , 2007, IEEE Transactions on Signal Processing.

[44]  Lester Melie-García,et al.  Estimating brain functional connectivity with sparse multivariate autoregression , 2005, Philosophical Transactions of the Royal Society B: Biological Sciences.

[45]  S. Mallat,et al.  Adaptive greedy approximations , 1997 .

[46]  M. Rudelson Random Vectors in the Isotropic Position , 1996, math/9608208.

[47]  Emmanuel J. Candès,et al.  Decoding by linear programming , 2005, IEEE Transactions on Information Theory.

[48]  Ting Sun,et al.  Single-pixel imaging via compressive sampling , 2008, IEEE Signal Process. Mag..

[49]  Namrata Vaswani,et al.  Kalman filtered Compressed Sensing , 2008, 2008 15th IEEE International Conference on Image Processing.

[50]  J. Mendel Lessons in Estimation Theory for Signal Processing, Communications, and Control , 1995 .

[51]  R. DeVore,et al.  A Simple Proof of the Restricted Isometry Property for Random Matrices , 2008 .

[52]  Michael A. Saunders,et al.  Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..