Iteration complexity of feasible descent methods for convex optimization
暂无分享,去创建一个
[1] A. Hoffman. On approximate solutions of systems of linear inequalities , 1952 .
[2] Clifford Hildreth,et al. A quadratic programming procedure , 1957 .
[3] S. M. Robinson. Bounds for error in the solution set of a perturbed linear program , 1973 .
[4] Jean-Philippe Vial,et al. Strong and Weak Convexity of Sets and Functions , 1983, Math. Oper. Res..
[5] Gene H. Golub,et al. Matrix computations , 1983 .
[6] D. Bertsekas,et al. TWO-METRIC PROJECTION METHODS FOR CONSTRAINED OPTIMIZATION* , 1984 .
[7] Jong-Shi Pang,et al. A Posteriori Error Bounds for the Linearly-Constrained Variational Inequality Problem , 1987, Math. Oper. Res..
[8] O. Mangasarian,et al. Lipschitz continuity of solutions of linear inequalities, programs and complementarity problems , 1987 .
[9] P. Tseng,et al. On the linear convergence of descent methods for convex essentially smooth minimization , 1992 .
[10] Bernhard E. Boser,et al. A training algorithm for optimal margin classifiers , 1992, COLT '92.
[11] P. Tseng,et al. On the convergence of the coordinate descent method for convex differentiable minimization , 1992 .
[12] Z.-Q. Luo,et al. Error bounds and convergence analysis of feasible descent methods: a general approach , 1993, Ann. Oper. Res..
[13] Wu Li. Sharp Lipschitz Constants for Basic Optimal Solutions and Basic Feasible Solutions of Linear Programs , 1994 .
[14] John C. Platt,et al. Fast training of support vector machines using sequential minimal optimization, advances in kernel methods , 1999 .
[15] Vladimir N. Vapnik,et al. The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.
[16] J. Hiriart-Urruty,et al. Fundamentals of Convex Analysis , 2004 .
[17] Chih-Jen Lin,et al. Trust region Newton methods for large-scale logistic regression , 2007, ICML '07.
[18] Chih-Jen Lin,et al. A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.
[19] Chih-Jen Lin,et al. Coordinate Descent Method for Large-scale L2-loss Linear Support Vector Machines , 2008, J. Mach. Learn. Res..
[20] P. Tseng,et al. Block-Coordinate Gradient Descent Method for Linearly Constrained Nonsmooth Separable Optimization , 2009 .
[21] Cho-Jui Hsieh,et al. Coordinate Descent Method for Large-scale L 2-loss Linear SVM , 2008 .
[22] Chih-Jen Lin,et al. Trust Region Newton Method for Logistic Regression , 2008, J. Mach. Learn. Res..
[23] Ambuj Tewari,et al. Stochastic methods for l1 regularized loss minimization , 2009, ICML '09.
[24] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[25] Stephen J. Wright. Accelerated Block-coordinate Relaxation for Regularized Optimization , 2012, SIAM J. Optim..
[26] Chia-Hua Ho,et al. Large-scale linear support vector regression , 2012, J. Mach. Learn. Res..
[27] Amir Beck,et al. On the Convergence of Block Coordinate Descent Type Methods , 2013, SIAM J. Optim..
[28] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[29] Ambuj Tewari,et al. On the Nonasymptotic Convergence of Cyclic Coordinate Descent Methods , 2013, SIAM J. Optim..
[30] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[31] Peter Richtárik,et al. Inexact Coordinate Descent: Complexity and Preconditioning , 2013, J. Optim. Theory Appl..
[32] Leon Hirsch,et al. Fundamentals Of Convex Analysis , 2016 .
[33] Tong Zhang,et al. Accelerated proximal stochastic dual coordinate ascent for regularized loss minimization , 2013, Mathematical Programming.