Efficient and Consistent Robust Time Series Analysis

We study the problem of robust time series analysis under the standard auto-regressive (AR) time series model in the presence of arbitrary outliers. We devise an efficient hard thresholding based algorithm which can obtain a consistent estimate of the optimal AR model despite a large fraction of the time series points being corrupted. Our algorithm alternately estimates the corrupted set of points and the model parameters, and is inspired by recent advances in robust regression and hard-thresholding methods. However, a direct application of existing techniques is hindered by a critical difference in the time-series domain: each point is correlated with all previous points rendering existing tools inapplicable directly. We show how to overcome this hurdle using novel proof techniques. Using our techniques, we are also able to provide the first efficient and provably consistent estimator for the robust regression problem where a standard linear observation model with white additive noise is corrupted arbitrarily. We illustrate our methods on synthetic datasets and show that our methods indeed are able to consistently recover the optimal parameters despite a large fraction of points being corrupted.

[1]  Trac D. Tran,et al.  Exact Recoverability From Dense Corrupted Observations via $\ell _{1}$-Minimization , 2011, IEEE Transactions on Information Theory.

[2]  Roman Vershynin,et al.  Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.

[3]  V. Yohai,et al.  Robust Estimation for ARMA models , 2009, 0904.0106.

[4]  R. Martin Robust Estimation for Time Series Autoregressions , 1979 .

[5]  Christophe Croux,et al.  Robust Estimation of the Vector Autoregressive Model by a Least Trimmed Squares Procedure , 2008 .

[6]  Ohad Shamir A Variant of Azuma's Inequality for Martingales with Subgaussian Tails , 2011, ArXiv.

[7]  Rudolf Dutter,et al.  Robust time series analysis: a survey , 1987, Kybernetika.

[8]  Elad Hazan,et al.  Online Time Series Prediction with Missing Data , 2015, ICML.

[9]  Rahul Garg,et al.  Gradient descent with sparsification: an iterative algorithm for sparse recovery with restricted isometry property , 2009, ICML '09.

[10]  Prateek Jain,et al.  On Iterative Hard Thresholding Methods for High-dimensional M-Estimation , 2014, NIPS.

[11]  M. Rudelson,et al.  Hanson-Wright inequality and sub-gaussian concentration , 2013 .

[12]  P. Rousseeuw Least Median of Squares Regression , 1984 .

[13]  Prateek Jain,et al.  Robust Regression via Hard Thresholding , 2015, NIPS.

[14]  John Wright,et al.  Dense Error Correction Via $\ell^1$-Minimization , 2010, IEEE Transactions on Information Theory.

[15]  John Wright,et al.  Dense Error Correction via L1-Minimization , 2008, 0809.0199.

[16]  A. Banerjee,et al.  Estimating Structured Vector Autoregressive Models , 2016, ICML.

[17]  G. Michailidis,et al.  Regularized estimation in sparse high-dimensional time series models , 2013, 1311.4175.

[18]  Jan Ámos Vísek,et al.  The least trimmed squares. Part I: Consistency , 2006, Kybernetika.

[19]  Shie Mannor,et al.  Robust Sparse Regression under Adversarial Corruption , 2013, ICML.

[20]  MaYi,et al.  Dense error correction via l1-minimization , 2010 .

[21]  V. Yohai,et al.  Robust Statistics: Theory and Methods , 2006 .

[22]  Mike E. Davies,et al.  Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.