Changepoint Detection by the Quantile LASSO Method

A simultaneous change-point detection and estimation in a piece-wise constant model is a common task in modern statistics. If, in addition, the whole estimation can be performed automatically, in just one single step without going through any hypothesis tests for non-identifiable models, or unwieldy classical a-posterior methods, it becomes an interesting, but also challenging idea. In this paper we introduce the estimation method based on the quantile LASSO approach. Unlike standard LASSO approaches, our method does not rely on typical assumptions usually required for the model errors, such as sub-Gaussian or Normal distribution. The proposed quantile LASSO method can effectively handle heavy-tailed random error distributions, and, in general, it offers a more complex view of the data as one can obtain any conditional quantile of the target distribution, not just the conditional mean. It is proved that under some reasonable assumptions the number of change-points is not underestimated with probability tenting to one, and, in addition, when the number of change-points is estimated correctly, the change-point estimates provided by the quantile LASSO are consistent. Numerical simulations are used to demonstrate these results and to illustrate the empirical performance robust favor of the proposed quantile LASSO method.

[1]  M. Hušková,et al.  Discontinuities in robust nonparametric regression with α-mixing dependence , 2017 .

[2]  Gabriela Ciuperca Adaptive LASSO model selection in a multiphase quantile regression , 2013, 1309.1262.

[3]  R. Tibshirani,et al.  Regression shrinkage and selection via the lasso: a retrospective , 2011 .

[4]  D. Donoho,et al.  Atomic Decomposition by Basis Pursuit , 2001 .

[5]  R. Tibshirani,et al.  Spatial smoothing and hot spot detection for CGH data using the fused lasso. , 2008, Biostatistics.

[6]  A. Munk,et al.  Multiscale change point inference , 2013, 1301.7212.

[7]  R. Tibshirani,et al.  Sparsity and smoothness via the fused lasso , 2005 .

[8]  Peng Zhao,et al.  On Model Selection Consistency of Lasso , 2006, J. Mach. Learn. Res..

[9]  Brian S. Yandell,et al.  A local polynomial jump-detection algorithm in nonparametric regression , 1998 .

[10]  Max G'Sell,et al.  Exact post-selection inference for the generalized lasso path , 2018 .

[11]  J. Kiefer,et al.  Asymptotic Minimax Character of the Sample Distribution Function and of the Classical Multinomial Estimator , 1956 .

[12]  Z. Harchaoui,et al.  Multiple Change-Point Estimation With a Total Variation Penalty , 2010 .

[13]  Testing structural changes in panel data with small fixed panel size and bootstrap , 2015, 1509.01291.

[14]  Matúš Maciak,et al.  Regularization techniques in joinpoint regression , 2016 .

[15]  V. Liebscher,et al.  Consistencies and rates of convergence of jump-penalized least squares estimators , 2009, 0902.4838.

[16]  Keith Knight,et al.  Limiting distributions for $L\sb 1$ regression estimators under general conditions , 1998 .

[17]  Victor Chernozhukov,et al.  Quantile regression , 2019, Journal of Econometrics.

[18]  H. Zou The Adaptive Lasso and Its Oracle Properties , 2006 .

[19]  Junhui Qian,et al.  SHRINKAGE ESTIMATION OF REGRESSION MODELS WITH MULTIPLE STRUCTURAL CHANGES , 2015, Econometric Theory.

[20]  Q. Shao,et al.  A general bahadur representation of M-estimators and its application to linear regression with nonstochastic designs , 1996 .

[21]  R. Tibshirani Adaptive piecewise polynomial estimation via trend filtering , 2013, 1304.2986.

[22]  Irène Gijbels,et al.  Nonparametric simultaneous testing for structural breaks , 2008 .

[23]  A. Rinaldo,et al.  Approximate Recovery in Changepoint Problems, from $\ell_2$ Estimation Error Rates , 2016, 1606.06746.

[24]  N. Meinshausen,et al.  High-dimensional graphs and variable selection with the Lasso , 2006, math/0608017.

[25]  Myung Hwan Seo,et al.  The lasso for high dimensional regression with a possible change point , 2012, Journal of the Royal Statistical Society. Series B, Statistical methodology.

[26]  P. Kokoszka,et al.  Change-Point Detection With Non-Parametric Regression , 2002 .

[27]  Gabriela Ciuperca,et al.  Change‐point detection in a linear model by adaptive fused quantile method , 2019, Scandinavian Journal of Statistics.

[28]  R. Tibshirani,et al.  Exact Post-Selection Inference for Changepoint Detection and Other Generalized Lasso Problems , 2016, 1606.03552.

[29]  S. Geer,et al.  Locally adaptive regression splines , 1997 .

[30]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .

[31]  Tests for continuity of regression functions , 2007 .

[32]  I. Gijbels,et al.  Curve Fitting Under Jump and Peak Irregularities Using Local Linear Regression , 2011 .

[33]  Peter Bühlmann Regression shrinkage and selection via the Lasso: a retrospective (Robert Tibshirani): Comments on the presentation , 2011 .

[34]  L. Horváth,et al.  Limit Theorems in Change-Point Analysis , 1997 .

[35]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[36]  Jianqing Fan,et al.  ADAPTIVE ROBUST VARIABLE SELECTION. , 2012, Annals of statistics.

[37]  Binbing Yu,et al.  SELECTING THE NUMBER OF CHANGE-POINTS IN SEGMENTED LINE REGRESSION. , 2009, Statistica Sinica.

[38]  Gabriela Ciuperca Model selection by LASSO methods in a change-point model , 2011 .

[39]  R. Koenker Quantile Regression: Name Index , 2005 .