Bolasso: model consistent Lasso estimation through the bootstrap

We consider the least-square linear regression problem with regularization by the l1-norm, a problem usually referred to as the Lasso. In this paper, we present a detailed asymptotic analysis of model consistency of the Lasso. For various decays of the regularization parameter, we compute asymptotic equivalents of the probability of correct model selection (i.e., variable selection). For a specific rate decay, we show that the Lasso selects all the variables that should enter the model with probability tending to one exponentially fast, while it selects all other variables with strictly positive probability. We show that this property implies that if we run the Lasso for several bootstrapped replications of a given sample, then intersecting the supports of the Lasso bootstrap estimates leads to consistent model selection. This novel variable selection algorithm, referred to as the Bolasso, is compared favorably to other linear regression methods on synthetic data and datasets from the UCI machine learning repository.

[1]  Wenjiang J. Fu,et al.  Asymptotics for lasso-type estimators , 2000 .

[2]  P. Bühlmann Boosting for high-dimensional linear models , 2006 .

[3]  Francis R. Bach,et al.  Consistency of the group Lasso and multiple kernel learning , 2007, J. Mach. Learn. Res..

[4]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[5]  M. Kenward,et al.  An Introduction to the Bootstrap , 2007 .

[6]  Karim Lounici Sup-norm convergence rate and sign concentration property of Lasso and Dantzig estimators , 2008, 0801.4610.

[7]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[8]  V. Bentkus On the dependence of the Berry–Esseen bound on dimension , 2003 .

[9]  M. Yuan,et al.  On the non‐negative garrotte estimator , 2007 .

[10]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[11]  L. Breiman Heuristics of instability and stabilization in model selection , 1996 .

[12]  Peter Buhlmann Boosting for high-dimensional linear models , 2006, math/0606789.

[13]  N. Meinshausen,et al.  LASSO-TYPE RECOVERY OF SPARSE REPRESENTATIONS FOR HIGH-DIMENSIONAL DATA , 2008, 0806.0145.

[14]  H. Zou The Adaptive Lasso and Its Oracle Properties , 2006 .

[15]  Martin J. Wainwright,et al.  Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$ -Constrained Quadratic Programming (Lasso) , 2009, IEEE Transactions on Information Theory.

[16]  Peng Zhao,et al.  On Model Selection Consistency of Lasso , 2006, J. Mach. Learn. Res..