Valid Post-Selection and Post-Regularization Inference: An Elementary, General Approach

Here we present an expository, general analysis of valid post-selection or post-regularization inference about a low-dimensional target parameter, α, in the presence of a very high-dimensional nuisance parameter, η, which is estimated using modern selection or regularization methods. Our analysis relies on high-level, easy-to-interpret conditions that allow one to clearly see the structures needed for achieving valid post-regularization inference. Simple, readily verifiable sufficient conditions are provided for a class of affine-quadratic models. We rely on asymptotic statements which dramatically simplifies theoretical statements and helps highlight the structure of the problem. We focus our discussion on estimation and inference procedures based on using the empirical analog of theoretical equations M(α, η) = 0 which identify α. Within this structure, we show that setting up such equations in a manner such that the orthogonality/immunization condition ∂ηM (α, η) = 0 at the true parameter values is satisfied, coupled with plausible conditions on the smoothness of M and the quality of the estimator ηˆ, guarantees that inference for the main parameter α based on testing or point estimation methods discussed below will be regular despite selection or regularization biases occurring in estimation of η. In particular, the estimator of α will often be uniformly consistent at the root-n rate and uniformly asymptotically normal even though estimators ηˆ will generally not be asymptotically linear and regular. The uniformity holds over large classes of models that do not impose highly implausible “beta-min” conditions. We also show that inference can be carried out by inverting tests formed from Neyman’s C(α) (orthogonal score) statistics. As an application and an illustration of these ideas, we provide an analysis of post-selection inference in the linear models with many regressors and many instruments. We conclude with a review of other developments in post-selection inference and argue that many of the developments can be viewed as special cases of the general framework of orthogonalized estimating equations.

[1]  M. Farrell Robust Inference on Average Treatment Effects with Possibly More Covariates than Observations , 2013, 1309.4686.

[2]  P. J. Huber The behavior of maximum likelihood estimates under nonstandard conditions , 1967 .

[3]  A. Belloni,et al.  Honest Confidence Regions for Logistic Regression with a Large Number of Controls , 2013 .

[4]  Martin Spindler,et al.  Post-Selection and Post-Regularization Inference in Linear Models with Many Controls and Instruments , 2015 .

[5]  Victor Chernozhukov,et al.  Inference on Treatment Effects after Selection Amongst High-Dimensional Controls , 2011 .

[6]  G. Chamberlain Asymptotic efficiency in estimation with conditional moment restrictions , 1987 .

[7]  H. Leeb,et al.  Sparse Estimators and the Oracle Property, or the Return of Hodges' Estimator , 2007, 0704.1466.

[8]  Prem S. Puri,et al.  On Optimal Asymptotic Tests of Composite Statistical Hypotheses , 1967 .

[9]  R. Tibshirani,et al.  Adaptive testing for the graphical lasso , 2013, 1307.4765.

[10]  Shuheng Zhou,et al.  25th Annual Conference on Learning Theory Reconstruction from Anisotropic Random Measurements , 2022 .

[11]  A. Belloni,et al.  Least Squares After Model Selection in High-Dimensional Sparse Models , 2009, 1001.0188.

[12]  Cun-Hui Zhang,et al.  Confidence intervals for low dimensional parameters in high dimensional linear models , 2011, 1110.2563.

[13]  Robert Tibshirani,et al.  Post-selection adaptive inference for Least Angle Regression and the Lasso , 2014 .

[14]  Sara van de Geer,et al.  Confidence sets in sparse regression , 2012, 1209.1508.

[15]  D. Pollard,et al.  Simulation and the Asymptotics of Optimization Estimators , 1989 .

[16]  Kengo Kato,et al.  Uniform post selection inference for LAD regression models , 2013 .

[17]  Sara van de Geer,et al.  Statistics for High-Dimensional Data: Methods, Theory and Applications , 2011 .

[18]  A. Belloni,et al.  Program evaluation with high-dimensional data , 2013 .

[19]  Han Liu,et al.  SPARC: Optimal Estimation and Asymptotic Inference under Semiparametric Sparsity , 2014 .

[20]  S. Geer,et al.  On asymptotically optimal confidence regions and tests for high-dimensional models , 2013, 1303.0518.

[21]  Jianqing Fan,et al.  Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .

[22]  A. Buja,et al.  Valid post-selection inference , 2013, 1306.1059.

[23]  Dennis L. Sun,et al.  Exact post-selection inference with the lasso , 2013 .

[24]  Kengo Kato,et al.  Gaussian approximations and multiplier bootstrap for maxima of sums of high-dimensional random vectors , 2013 .

[25]  Dennis L. Sun,et al.  Optimal Inference After Model Selection , 2014, 1410.2597.

[26]  Victor Chernozhukov,et al.  High Dimensional Sparse Econometric Models : An , 2011 .

[27]  A. Belloni,et al.  Inference for High-Dimensional Sparse Econometric Models , 2011, 1201.0220.

[28]  Joshua R. Loftus,et al.  A significance test for forward stepwise model selection , 2014, 1405.3920.

[29]  A. Belloni,et al.  SPARSE MODELS AND METHODS FOR OPTIMAL INSTRUMENTS WITH AN APPLICATION TO EMINENT DOMAIN , 2012 .

[30]  J. Robins,et al.  Semiparametric Efficiency in Multivariate Regression Models with Missing Data , 1995 .

[31]  Adel Javanmard,et al.  Confidence intervals and hypothesis testing for high-dimensional regression , 2013, J. Mach. Learn. Res..

[32]  A. Tsybakov,et al.  High-dimensional instrumental variables regression and confidence sets -- v2/2012 , 2018, 1812.11330.

[33]  Jianqing Fan,et al.  A Selective Overview of Variable Selection in High Dimensional Feature Space. , 2009, Statistica Sinica.

[34]  Christian Hansen,et al.  Inference in High-Dimensional Panel Models With an Application to Gun Control , 2014, 1411.6507.

[35]  Ali Shojaie,et al.  Inference in High Dimensions with the Penalized Score Test , 2014, 1401.2678.

[36]  Jonathan E. Taylor,et al.  Exact Post Model Selection Inference for Marginal Screening , 2014, NIPS.

[37]  Ulf Grenander,et al.  Probability And Statistics The Harald Cramer Volume , 1962 .

[38]  R. Tibshirani,et al.  A SIGNIFICANCE TEST FOR THE LASSO. , 2013, Annals of statistics.

[39]  Xiaohong Chen,et al.  Estimation of Semiparametric Models When the Criterion Function is Not Smooth , 2002 .