Conditional SIRS for nonparametric and semiparametric models by marginal empirical likelihood

Dimension reduction is a crucial issue for high-dimensional data analysis. When the correlation among the variables is strong, the original SIRS (Zhu et al. in J Am Stat Assoc 106(496):1464–1475, 2011) may lose efficiency. Under high-dimensional setting, eliminating the bad influence caused by the correlation has become an important issue. Aiming at this issue, we propose a feature screening approach by combining the marginal empirical likelihood with the conditional SIRS. Based on a centralized SIRS, the correlation among the variables is significantly reduced and consequently, the related empirical likelihood is improved remarkably. Moreover, our method is model-free due to the properties of SIRS and empirical likelihood. The proposed method can select important predictors directly without parameter estimation, implying that the method is computationally simple. Under some general conditions, the proposed marginal empirical likelihood ratio is self-studentized. The simulation study shows that compared with other unconditional and conditional methods, our method is competitive and has a great superiority.

[1]  Chen Xu,et al.  The Sparse MLE for Ultrahigh-Dimensional Feature Screening , 2014, Journal of the American Statistical Association.

[2]  J. Mesirov,et al.  Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. , 1999, Science.

[3]  Hansheng Wang,et al.  Robust Regression Shrinkage and Consistent Variable Selection Through the LAD-Lasso , 2007 .

[4]  Jianqing Fan,et al.  Conditional Sure Independence Screening , 2012, Journal of the American Statistical Association.

[5]  Qinqin Hu,et al.  Conditional sure independence screening by conditional marginal empirical likelihood , 2015, Annals of the Institute of Statistical Mathematics.

[6]  Yichao Wu,et al.  LOCAL INDEPENDENCE FEATURE SCREENING FOR NONPARAMETRIC AND SEMIPARAMETRIC MODELS BY MARGINAL EMPIRICAL LIKELIHOOD. , 2015, Annals of statistics.

[7]  Runze Li,et al.  Feature Selection for Varying Coefficient Models With Ultrahigh-Dimensional Covariates , 2014, Journal of the American Statistical Association.

[8]  Lixing Zhu,et al.  Nonparametric feature screening , 2013, Comput. Stat. Data Anal..

[9]  Jianqing Fan,et al.  Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .

[10]  Hulin Wu,et al.  Variable Selection for Sparse High-Dimensional Nonlinear Regression Models by Combining Nonnegative Garrote and Sure Independence Screening. , 2014, Statistica Sinica.

[11]  Runze Li,et al.  Feature Screening via Distance Correlation Learning , 2012, Journal of the American Statistical Association.

[12]  Lan Wang,et al.  Quantile-adaptive model-free variable screening for high-dimensional heterogeneous data , 2013, 1304.2186.

[13]  H. Zou The Adaptive Lasso and Its Oracle Properties , 2006 .

[14]  L. Breiman Better subset regression using the nonnegative garrote , 1995 .

[15]  A. Owen Empirical likelihood ratio confidence intervals for a single functional , 1988 .

[16]  A. Owen Empirical Likelihood Ratio Confidence Regions , 1990 .

[17]  Bovas Abraham,et al.  Adjusted Empirical Likelihood and its Properties , 2008 .

[18]  W Y Zhang,et al.  Discussion on `Sure independence screening for ultra-high dimensional feature space' by Fan, J and Lv, J. , 2008 .

[19]  Jianqing Fan,et al.  Sure independence screening in generalized linear models with NP-dimensionality , 2009, The Annals of Statistics.

[20]  Jianqing Fan,et al.  Sure independence screening for ultrahigh dimensional feature space , 2006, math/0612857.

[21]  G. Tian,et al.  Adaptive group Lasso for high-dimensional generalized linear models , 2019 .

[22]  Jianqing Fan,et al.  Nonparametric Independence Screening in Sparse Ultra-High-Dimensional Varying Coefficient Models , 2014, Journal of the American Statistical Association.

[23]  Terence Tao,et al.  The Dantzig selector: Statistical estimation when P is much larger than n , 2005, math/0506081.

[24]  Jun Lu,et al.  Model-free conditional screening via conditional distance correlation , 2020 .

[25]  H. Zou,et al.  Composite quantile regression and the oracle Model Selection Theory , 2008, 0806.2905.

[26]  Jing Sun,et al.  Adaptive conditional feature screening , 2016, Comput. Stat. Data Anal..

[27]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[28]  J. Friedman,et al.  A Statistical View of Some Chemometrics Regression Tools , 1993 .

[29]  Yichao Wu,et al.  MARGINAL EMPIRICAL LIKELIHOOD AND SURE INDEPENDENCE FEATURE SCREENING. , 2013, Annals of statistics.

[30]  Runze Li,et al.  Model-Free Feature Screening for Ultrahigh-Dimensional Data , 2011, Journal of the American Statistical Association.