On the Asymptotic Properties of The Group Lasso Estimator in Least Squares Problems
暂无分享,去创建一个
[1] Karim Lounici. Sup-norm convergence rate and sign concentration property of Lasso and Dantzig estimators , 2008, 0801.4610.
[2] M. Yuan,et al. Model selection and estimation in regression with grouped variables , 2006 .
[3] Francis R. Bach,et al. Consistency of the group Lasso and multiple kernel learning , 2007, J. Mach. Learn. Res..
[4] P. Bühlmann,et al. Sparse Contingency Tables and High-Dimensional Log-Linear Models for Alternative Splicing in Full-Length cDNA Libraries , 2006 .
[5] Martin J. Wainwright,et al. Information-theoretic limits on sparsity recovery in the high-dimensional and noisy setting , 2009, IEEE Trans. Inf. Theory.
[6] Tong Zhang. Some sharp performance bounds for least squares regression with L1 regularization , 2009, 0908.2869.
[7] Martin J. Wainwright,et al. Sharp thresholds for high-dimensional and noisy recovery of sparsity , 2006, ArXiv.
[8] P. Bühlmann,et al. The group lasso for logistic regression , 2008 .
[9] Joel A. Tropp,et al. Algorithms for simultaneous sparse approximation. Part I: Greedy pursuit , 2006, Signal Process..
[10] V. Koltchinskii. Sparsity in penalized empirical risk minimization , 2009 .
[11] Jon A. Wellner,et al. Weak Convergence and Empirical Processes: With Applications to Statistics , 1996 .
[12] Cun-Hui Zhang,et al. The sparsity and bias of the Lasso selection in high-dimensional linear regression , 2008, 0808.0967.
[13] S. Portnoy. Asymptotic Behavior of Likelihood Methods for Exponential Families when the Number of Parameters Tends to Infinity , 1988 .
[14] Y. Ritov,et al. Persistence in high-dimensional linear predictor selection and the virtue of overparametrization , 2004 .
[15] H. Zou. The Adaptive Lasso and Its Oracle Properties , 2006 .
[16] A. Tsybakov,et al. Sparsity oracle inequalities for the Lasso , 2007, 0705.3308.
[17] Larry A. Wasserman,et al. Compressed Regression , 2007, NIPS.
[18] A. Tsybakov,et al. Aggregation for Gaussian regression , 2007, 0710.3654.
[19] N. Meinshausen,et al. High-dimensional graphs and variable selection with the Lasso , 2006, math/0608017.
[20] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[21] J. Tropp. Algorithms for simultaneous sparse approximation. Part II: Convex relaxation , 2006, Signal Process..
[22] P. Bickel,et al. SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR , 2008, 0801.1095.
[23] Ji Zhu,et al. A ug 2 01 0 Group Variable Selection via a Hierarchical Lasso and Its Oracle Property Nengfeng Zhou Consumer Credit Risk Solutions Bank of America Charlotte , NC 28255 , 2010 .
[24] Alessandro Rinaldo,et al. Computing Maximum Likelihood Estimates in Log-Linear Models , 2006 .
[25] M. Talagrand,et al. Probability in Banach Spaces: Isoperimetry and Processes , 1991 .
[26] Jianqing Fan,et al. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .
[27] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[28] Wenjiang J. Fu,et al. Asymptotics for lasso-type estimators , 2000 .
[29] M. R. Osborne,et al. On the LASSO and its Dual , 2000 .
[30] E. Greenshtein. Best subset selection, persistence in high-dimensional statistical learning and optimization under l1 constraint , 2006, math/0702684.
[31] J. Lafferty,et al. Sparse additive models , 2007, 0711.4555.
[32] Jianqing Fan,et al. Nonconcave penalized likelihood with a diverging number of parameters , 2004, math/0406466.
[33] P. Massart,et al. Concentration inequalities and model selection , 2007 .
[34] Peng Zhao,et al. On Model Selection Consistency of Lasso , 2006, J. Mach. Learn. Res..
[35] P. Zhao,et al. Grouped and Hierarchical Model Selection through Composite Absolute Penalties , 2007 .
[36] Martin J. Wainwright,et al. Information-Theoretic Limits on Sparsity Recovery in the High-Dimensional and Noisy Setting , 2007, IEEE Transactions on Information Theory.