Stochastic Coordinate Descent Methods for Regularized Smooth and Nonsmooth Losses
暂无分享,去创建一个
Qing Tao | Kang Kong | Gao-wei Wu | Dejun Chu | Qing Tao | Gao-wei Wu | Dejun Chu | Kang-Kook Kong
[1] Chih-Jen Lin,et al. A Comparison of Optimization Methods and Software for Large-scale L1-regularized Linear Classification , 2010, J. Mach. Learn. Res..
[2] Jun Liu,et al. Efficient Euclidean projections in linear time , 2009, ICML '09.
[3] Chih-Jen Lin,et al. Trust Region Newton Method for Logistic Regression , 2008, J. Mach. Learn. Res..
[4] Thorsten Joachims,et al. Training linear SVMs in linear time , 2006, KDD '06.
[5] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[6] Trevor Hastie,et al. Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.
[7] Chih-Jen Lin,et al. Coordinate Descent Method for Large-scale L2-loss Linear Support Vector Machines , 2008, J. Mach. Learn. Res..
[8] Y. Nesterov. Gradient methods for minimizing composite objective function , 2007 .
[9] Yoram Singer,et al. Pegasos: primal estimated sub-gradient solver for SVM , 2007, ICML '07.
[10] Lin Xiao,et al. Dual Averaging Methods for Regularized Stochastic Learning and Online Optimization , 2009, J. Mach. Learn. Res..
[11] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[12] Yurii Nesterov,et al. Primal-dual subgradient methods for convex problems , 2005, Math. Program..
[13] Yurii Nesterov,et al. How to advance in Structural Convex Optimization , 2008 .
[14] Y. Nesterov. A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .
[15] Yoram Singer,et al. Pegasos: primal estimated sub-gradient solver for SVM , 2011, Math. Program..
[16] Yoram Singer,et al. Efficient projections onto the l1-ball for learning in high dimensions , 2008, ICML '08.
[17] Stephen P. Boyd,et al. An Interior-Point Method for Large-Scale l1-Regularized Logistic Regression , 2007, J. Mach. Learn. Res..
[18] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[19] Chih-Jen Lin,et al. A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.
[20] David R. Musicant,et al. Successive overrelaxation for support vector machines , 1999, IEEE Trans. Neural Networks.
[21] Ambuj Tewari,et al. Stochastic methods for l1 regularized loss minimization , 2009, ICML '09.
[22] Ambuj Tewari,et al. On the Finite Time Convergence of Cyclic Coordinate Descent Methods , 2010, ArXiv.
[23] Olvi L. Mangasarian,et al. A finite newton method for classification , 2002, Optim. Methods Softw..
[24] Yurii Nesterov,et al. Smooth minimization of non-smooth functions , 2005, Math. Program..
[25] Paul Tseng,et al. A coordinate gradient descent method for nonsmooth separable minimization , 2008, Math. Program..
[26] Zhengya Sun,et al. Developing Learning Algorithms via Optimized Discretization of Continuous Dynamical Systems , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).
[27] Kim-Chuan Toh,et al. A coordinate gradient descent method for ℓ1-regularized convex minimization , 2011, Comput. Optim. Appl..