Sparse and Low-bias Estimation of High Dimensional Vector Autoregressive Models

Vector autoregressive (VAR) models are widely used for causal discovery and forecasting in multivariate time series analysis. In the high-dimensional setting, which is increasingly common in fields such as neuroscience and econometrics, model parameters are inferred by L1-regularized maximum likelihood (RML). A well-known feature of RML inference is that in general the technique produces a trade-off between sparsity and bias that depends on the choice of the regularization hyperparameter. In the context of multivariate time series analysis, sparse estimates are favorable for causal discovery and low-bias estimates are favorable for forecasting. However, owing to a paucity of research on hyperparameter selection methods, practitioners must rely on ad-hoc methods such as cross-validation (or manual tuning). The particular balance that such approaches achieve between the two goals — causal discovery and forecasting — is poorly understood. Our paper investigates this behavior and proposes a method (UoI-VAR) that achieves a better balance between sparsity and bias when the underlying causal influences are in fact sparse. We demonstrate through simulation that RML with a hyperparameter selected by cross-validation tends to overfit, producing relatively dense estimates. We further demonstrate that UoI-VAR much more effectively approximates the correct sparsity pattern with only a minor compromise in model fit, particularly so for larger data dimensions, and that the estimates produced by UoI-VAR exhibit less bias. We conclude that our method achieves improved performance especially well-suited to applications involving simultaneous causal discovery and forecasting in high-dimensional settings.

[1]  Scott T. Grafton,et al.  Dynamic reconfiguration of human brain networks during learning , 2010, Proceedings of the National Academy of Sciences.

[2]  C. Gilligan,et al.  Thirteen challenges in modelling plant diseases. , 2015, Epidemics.

[3]  Sara van de Geer,et al.  Statistics for High-Dimensional Data , 2011 .

[4]  Umberto Triacca,et al.  Anthropogenic global warming hypothesis: testing its robustness by Granger causality analysis , 2013 .

[5]  P. Bühlmann,et al.  Block length selection in the bootstrap for time series , 1999 .

[6]  J. Fackler,et al.  An Application of Vector Time Series Techniques to Macroeconomic Forecasting , 1986 .

[7]  Patrick Danaher,et al.  The joint graphical lasso for inverse covariance estimation across multiple classes , 2011, Journal of the Royal Statistical Society. Series B, Statistical methodology.

[8]  Kristofer E. Bouchard,et al.  Union of Intersections ( UoI ) for interpretable data driven discovery and prediction in neuroscience , 2018 .

[9]  J. Stock,et al.  Forecasting Using Principal Components From a Large Number of Predictors , 2002 .

[10]  Rebecca Willett,et al.  Inferring high-dimensional poisson autoregressive models , 2016, 2016 IEEE Statistical Signal Processing Workshop (SSP).

[11]  H. Künsch The Jackknife and the Bootstrap for General Stationary Observations , 1989 .

[12]  Anuj Karpatne,et al.  Machine Learning for the Geosciences: Challenges and Opportunities , 2017, IEEE Transactions on Knowledge and Data Engineering.

[13]  Eero P. Simoncelli,et al.  Spatio-temporal correlations and visual signalling in a complete neuronal population , 2008, Nature.

[14]  N. Meinshausen,et al.  Stability selection , 2008, 0809.2932.

[15]  G. A. Miller,et al.  Comparison of different cortical connectivity estimators for high‐resolution EEG recordings , 2007, Human brain mapping.

[16]  E. Levina,et al.  Joint estimation of multiple graphical models. , 2011, Biometrika.

[17]  G. Michailidis,et al.  Regularized estimation in sparse high-dimensional time series models , 2013, 1311.4175.

[18]  Marco Lippi,et al.  The Generalized Dynamic Factor Model , 2002 .

[19]  Fang Han,et al.  A direct estimation of high dimensional stationary vector autoregressions , 2013, J. Mach. Learn. Res..

[20]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[21]  C. Sims MACROECONOMICS AND REALITY , 1977 .

[22]  W. Härdle,et al.  Bootstrap Methods for Time Series , 2003 .

[23]  R. Kass,et al.  Multiple neural spike train data analysis: state-of-the-art and future challenges , 2004, Nature Neuroscience.

[24]  Eric R. Ziegel,et al.  Analysis of Financial Time Series , 2002, Technometrics.

[25]  Regina Y. Liu Moving blocks jackknife and bootstrap capture weak dependence , 1992 .

[26]  P. Bickel,et al.  Large Vector Auto Regressions , 2011, 1106.3915.

[27]  R. Tibshirani,et al.  PATHWISE COORDINATE OPTIMIZATION , 2007, 0708.1485.

[28]  Han Liu,et al.  Joint estimation of multiple graphical models from high dimensional time series , 2013, Journal of the Royal Statistical Society. Series B, Statistical methodology.

[29]  Helmut Ltkepohl,et al.  New Introduction to Multiple Time Series Analysis , 2007 .

[30]  Lei Qi,et al.  Sparse High Dimensional Models in Economics. , 2011, Annual review of economics.

[31]  C. Koch,et al.  The origin of extracellular fields and currents — EEG, ECoG, LFP and spikes , 2012, Nature Reviews Neuroscience.

[32]  Kesheng Wu,et al.  UoI-NMF Cluster: A Robust Nonnegative Matrix Factorization Algorithm for Improved Parts-Based Decomposition and Reconstruction of Noisy Data , 2017, 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA).

[33]  Daniel A. McFarland,et al.  Sociology in the Era of Big Data: The Ascent of Forensic Social Science , 2015, The American Sociologist.