A SHRINKAGE PRINCIPLE FOR HEAVY-TAILED DATA: HIGH-DIMENSIONAL ROBUST LOW-RANK MATRIX RECOVERY.
暂无分享,去创建一个
[1] John Darzentas,et al. Problem Complexity and Method Efficiency in Optimization , 1983 .
[2] Dimitri P. Bertsekas,et al. On the Douglas—Rachford splitting method and the proximal point algorithm for maximal monotone operators , 1992, Math. Program..
[3] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[4] Michael A. Saunders,et al. Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..
[5] G. Reinsel,et al. Multivariate Reduced-Rank Regression: Theory and Applications , 1998 .
[6] R. Ibragimov,et al. Short Communications: On an Exact Constant for the Rosenthal Inequality , 1998 .
[7] Jianqing Fan,et al. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .
[8] J. Stock,et al. Macroeconomic Forecasting Using Diffusion Indexes , 2002 .
[9] Jianqing Fan,et al. Sure independence screening for ultrahigh dimensional feature space , 2006, math/0612857.
[10] N. Meinshausen,et al. High-dimensional graphs and variable selection with the Lasso , 2006, math/0608017.
[11] Emmanuel J. Candès,et al. Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.
[12] Mário A. T. Figueiredo,et al. Gradient Projection for Sparse Reconstruction: Application to Compressed Sensing and Other Inverse Problems , 2007, IEEE Journal of Selected Topics in Signal Processing.
[13] Terence Tao,et al. The Dantzig selector: Statistical estimation when P is much larger than n , 2005, math/0506081.
[14] H. Zou,et al. One-step Sparse Estimates in Nonconcave Penalized Likelihood Models. , 2008, Annals of statistics.
[15] E. Candès. The restricted isometry property and its implications for compressed sensing , 2008 .
[16] Martin J. Wainwright,et al. A unified framework for high-dimensional analysis of $M$-estimators with decomposable regularizers , 2009, NIPS.
[17] P. Bickel,et al. SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR , 2008, 0801.1095.
[18] Jianqing Fan,et al. Sparsistency and Rates of Convergence in Large Covariance Matrix Estimation. , 2007, Annals of statistics.
[19] P. Bickel,et al. Covariance regularization by thresholding , 2009, 0901.3079.
[20] Pablo A. Parrilo,et al. Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization , 2007, SIAM Rev..
[21] Cun-Hui Zhang. Nearly unbiased variable selection under minimax concave penalty , 2010, 1002.4734.
[22] V. Koltchinskii,et al. Nuclear norm penalization and optimal rates for noisy low rank matrix completion , 2010, 1011.6256.
[23] Eric P. Xing,et al. Tree-Guided Group Lasso for Multi-Task Regression with Structured Sparsity , 2009, ICML.
[24] A. Tsybakov,et al. Estimation of high-dimensional low-rank matrices , 2009, 0912.5338.
[25] Martin J. Wainwright,et al. Estimation of (near) low-rank matrices with noise and high-dimensional scaling , 2009, ICML.
[26] Emmanuel J. Candès,et al. Matrix Completion With Noise , 2009, Proceedings of the IEEE.
[27] O. Catoni. Challenging the empirical mean and empirical variance: a deviation study , 2010, 1009.2048.
[28] Martin J. Wainwright,et al. Minimax Rates of Estimation for High-Dimensional Linear Regression Over $\ell_q$ -Balls , 2009, IEEE Transactions on Information Theory.
[29] Weidong Liu,et al. Adaptive Thresholding for Sparse Covariance Matrix Estimation , 2011, 1102.2237.
[30] Emmanuel J. Candès,et al. Tight Oracle Inequalities for Low-Rank Matrix Recovery From a Minimal Number of Noisy Random Measurements , 2011, IEEE Transactions on Information Theory.
[31] Benjamin Recht,et al. A Simpler Approach to Matrix Completion , 2009, J. Mach. Learn. Res..
[32] Emmanuel J. Candès,et al. Exact Matrix Completion via Convex Optimization , 2008, Found. Comput. Math..
[33] Joel A. Tropp,et al. User-Friendly Tail Bounds for Sums of Random Matrices , 2010, Found. Comput. Math..
[34] Roman Vershynin,et al. Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.
[35] Harrison H. Zhou,et al. OPTIMAL RATES OF CONVERGENCE FOR SPARSE COVARIANCE MATRIX ESTIMATION , 2012, 1302.3030.
[36] Martin J. Wainwright,et al. Restricted strong convexity and weighted matrix completion: Optimal bounds with noise , 2010, J. Mach. Learn. Res..
[37] J. Tropp. User-Friendly Tools for Random Matrices: An Introduction , 2012 .
[38] Jianqing Fan,et al. Large covariance estimation by thresholding principal orthogonal complements , 2011, Journal of the Royal Statistical Society. Series B, Statistical methodology.
[39] Ziqiang Shi. Online Douglas-Rachford splitting method , 2013, ArXiv.
[40] Roberto Imbuzeiro Oliveira,et al. The lower tail of random quadratic forms with applications to ordinary least squares , 2013, ArXiv.
[41] Anru Zhang,et al. ROP: Matrix Recovery via Rank-One Projections , 2013, ArXiv.
[42] Daniel J. Hsu,et al. Approximate loss minimization with heavy tails , 2013, ArXiv.
[43] Andrea Montanari,et al. Accurate Prediction of Phase Transitions in Compressed Sensing via a Connection to Minimax Denoising , 2011, IEEE Transactions on Information Theory.
[44] Shuheng Zhou,et al. 25th Annual Conference on Learning Theory Reconstruction from Anisotropic Random Measurements , 2022 .
[45] Gábor Lugosi,et al. Concentration Inequalities - A Nonasymptotic Theory of Independence , 2013, Concentration Inequalities.
[46] Bingsheng He,et al. A Strictly Contractive Peaceman-Rachford Splitting Method for Convex Programming , 2014, SIAM J. Optim..
[47] Jianqing Fan,et al. Robust Estimation of High-Dimensional Mean Regression , 2014, 1410.2150.
[48] Anru Zhang,et al. Sparse Representation of a Polytope and Recovery of Sparse Signals and Low-Rank Matrices , 2013, IEEE Transactions on Information Theory.
[49] V. Koltchinskii,et al. Concentration inequalities and moment bounds for sample covariance operators , 2014, 1405.2468.
[50] Jianqing Fan,et al. ADAPTIVE ROBUST VARIABLE SELECTION. , 2012, Annals of statistics.
[51] Jianqing Fan,et al. High-Dimensional Statistics , 2014 .
[52] Shannon E. Ellis,et al. Transcriptome analysis reveals dysregulation of innate immune response genes and neuronal activity-dependent genes in autism , 2014, Nature Communications.
[53] Po-Ling Loh,et al. Statistical consistency and asymptotic normality for high-dimensional robust M-estimators , 2015, ArXiv.
[54] Joel A. Tropp,et al. An Introduction to Matrix Concentration Inequalities , 2015, Found. Trends Mach. Learn..
[55] The Elements of Financial Econometrics , 2015 .
[56] G. Lugosi,et al. Empirical risk minimization for heavy-tailed losses , 2014, 1406.2462.
[57] Stanislav Minsker. Geometric median and robust estimation in Banach spaces , 2013, 1308.1334.
[58] S. Frick,et al. Compressed Sensing , 2014, Computer Vision, A Reference Guide.
[59] Daniel M. Kane,et al. Robust Estimators in High Dimensions without the Computational Intractability , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).
[60] A. Dalalyan,et al. On the prediction loss of the lasso in the partially labeled setting , 2016, 1606.06179.
[61] Daniel J. Hsu,et al. Loss Minimization and Parameter Estimation with Heavy Tails , 2013, J. Mach. Learn. Res..
[62] Jianqing Fan,et al. LARGE COVARIANCE ESTIMATION THROUGH ELLIPTICAL FACTOR MODELS. , 2015, Annals of statistics.
[63] Stanislav Minsker. Sub-Gaussian estimators of the mean of a random matrix with heavy-tailed entries , 2016, The Annals of Statistics.
[64] Kim-Chuan Toh,et al. Max-norm optimization for robust matrix recovery , 2016, Mathematical Programming.
[65] Yuling Yan,et al. Inference and uncertainty quantification for noisy matrix completion , 2019, Proceedings of the National Academy of Sciences.
[66] Qiang Sun,et al. Adaptive Huber Regression , 2017, Journal of the American Statistical Association.
[67] Yuling Yan,et al. Noisy Matrix Completion: Understanding Statistical Guarantees for Convex Relaxation via Nonconvex Optimization , 2019, SIAM J. Optim..