Sparse Classification and Phase Transitions: A Discrete Optimization Perspective
暂无分享,去创建一个
[1] H. Cramér. Mathematical Methods of Statistics (PMS-9), Volume 9 , 1946 .
[2] D. Bertsekas. Projected Newton methods for optimization problems with simple constraints , 1981, 1981 20th IEEE Conference on Decision and Control including the Symposium on Adaptive Processes.
[3] Paul H. Calamai,et al. Projected gradient methods for linearly constrained problems , 1987, Math. Program..
[4] Ignacio E. Grossmann,et al. An outer-approximation algorithm for a class of mixed-integer nonlinear programs , 1987, Math. Program..
[5] Sven Leyffer,et al. Solving mixed integer nonlinear programs by outer approximation , 1994, Math. Program..
[6] Balas K. Natarajan,et al. Sparse Approximate Solutions to Linear Systems , 1995, SIAM J. Comput..
[7] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[8] Huan Liu,et al. Feature Selection for Classification , 1997, Intell. Data Anal..
[9] V. Vapnik. The Support Vector Method of Function Estimation , 1998 .
[10] Anthony Widjaja,et al. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond , 2003, IEEE Transactions on Neural Networks.
[11] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[12] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[13] Victoria Stodden,et al. Breakdown Point of Model Selection When the Number of Variables Exceeds the Number of Observations , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.
[14] S. Sathiya Keerthi,et al. A Fast Dual Algorithm for Kernel Logistic Regression , 2002, 2007 International Joint Conference on Neural Networks.
[15] Chih-Jen Lin,et al. A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.
[16] Gérard Cornuéjols,et al. An algorithmic framework for convex mixed integer nonlinear programs , 2008, Discret. Optim..
[17] Chih-Jen Lin,et al. Trust Region Newton Method for Logistic Regression , 2008, J. Mach. Learn. Res..
[18] Richard G. Baraniuk,et al. 1-Bit compressive sensing , 2008, 2008 42nd Annual Conference on Information Sciences and Systems.
[19] David L. Donoho,et al. Observed universality of phase transitions in high-dimensional geometry, with implications for modern data analysis and signal processing , 2009, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.
[20] Martin J. Wainwright,et al. Information-theoretic limits on sparsity recovery in the high-dimensional and noisy setting , 2009, IEEE Trans. Inf. Theory.
[21] Martin J. Wainwright,et al. Information-Theoretic Limits on Sparsity Recovery in the High-Dimensional and Noisy Setting , 2007, IEEE Transactions on Information Theory.
[22] Martin J. Wainwright,et al. Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$ -Constrained Quadratic Programming (Lasso) , 2009, IEEE Transactions on Information Theory.
[23] Francis R. Bach,et al. High-Dimensional Non-Linear Variable Selection through Hierarchical Kernel Learning , 2009, ArXiv.
[24] Trevor Hastie,et al. Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.
[25] Jianqing Fan,et al. Sure independence screening in generalized linear models with NP-dimensionality , 2009, The Annals of Statistics.
[26] Martin J. Wainwright,et al. Information-Theoretic Limits on Sparse Signal Recovery: Dense versus Sparse Measurement Matrices , 2008, IEEE Transactions on Information Theory.
[27] Robert D. Nowak,et al. Sample complexity for 1-bit compressed sensing and sparse classification , 2010, 2010 IEEE International Symposium on Information Theory.
[28] Chih-Jen Lin,et al. Dual coordinate descent methods for logistic regression and maximum entropy models , 2011, Machine Learning.
[29] R. Tibshirani,et al. Regression shrinkage and selection via the lasso: a retrospective , 2011 .
[30] Yoram Singer,et al. Pegasos: primal estimated sub-gradient solver for SVM , 2011, Math. Program..
[31] Yaniv Plan,et al. One‐Bit Compressed Sensing by Linear Programming , 2011, ArXiv.
[32] Julien Mairal,et al. Optimization with Sparsity-Inducing Penalties , 2011, Found. Trends Mach. Learn..
[33] Yaniv Plan,et al. Robust 1-bit Compressed Sensing and Sparse Logistic Regression: A Convex Programming Approach , 2012, IEEE Transactions on Information Theory.
[34] Laurent Jacques,et al. Robust 1-Bit Compressive Sensing via Binary Stable Embeddings of Sparse Vectors , 2011, IEEE Transactions on Information Theory.
[35] Aixia Guo,et al. Gene Selection for Cancer Classification using Support Vector Machines , 2014 .
[36] Emmanuel J. Candès,et al. False Discoveries Occur Early on the Lasso Path , 2015, ArXiv.
[37] Iain Dunning,et al. Computing in Operations Research Using Julia , 2013, INFORMS J. Comput..
[38] D. Bertsimas,et al. Best Subset Selection via a Modern Optimization Lens , 2015, 1507.03133.
[39] Martin J. Wainwright,et al. Sparse learning via Boolean relaxations , 2015, Mathematical Programming.
[40] Alan Edelman,et al. Julia: A Fresh Approach to Numerical Computing , 2014, SIAM Rev..
[41] Dimitris Bertsimas,et al. Logistic Regression: From Art to Science , 2017 .
[42] Iain Dunning,et al. JuMP: A Modeling Language for Mathematical Optimization , 2015, SIAM Rev..
[43] Bart P. G. Van Parys,et al. Sparse high-dimensional regression: Exact scalable algorithms and phase transitions , 2017, The Annals of Statistics.
[44] David Gamarnik,et al. High Dimensional Regression with Binary Coefficients. Estimating Squared Error and a Phase Transtition , 2017, COLT.