暂无分享,去创建一个
Yurii Nesterov | Filip Hanzely | Peter Richt'arik | Nikita Doikov | Y. Nesterov | Peter Richtárik | Filip Hanzely | N. Doikov
[1] Shumeet Baluja,et al. Advances in Neural Information Processing , 1994 .
[2] Yurii Nesterov,et al. Cubic regularization of Newton method and its global performance , 2006, Math. Program..
[3] Chih-Jen Lin,et al. Coordinate Descent Method for Large-scale L2-loss Linear Support Vector Machines , 2008, J. Mach. Learn. Res..
[4] Yurii Nesterov,et al. Accelerating the cubic regularization of Newton’s method on convex problems , 2005, Math. Program..
[5] Nicholas I. M. Gould,et al. On solving trust-region and other regularised subproblems in optimization , 2010, Math. Program. Comput..
[6] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[7] Jorge Nocedal,et al. On the Use of Stochastic Hessian Information in Optimization Methods for Machine Learning , 2011, SIAM J. Optim..
[8] Pradeep Ravikumar,et al. Nearest Neighbor based Greedy Coordinate Descent , 2011, NIPS.
[9] Nicholas I. M. Gould,et al. Adaptive cubic regularisation methods for unconstrained optimization. Part I: motivation, convergence and numerical results , 2011, Math. Program..
[10] Nicholas I. M. Gould,et al. Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity , 2011, Math. Program..
[11] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[12] Mark W. Schmidt,et al. A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets , 2012, NIPS.
[13] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[14] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[15] Christian L. Müller,et al. Optimization of Convex Functions with Random Pursuit , 2011, SIAM J. Optim..
[16] Renato D. C. Monteiro,et al. An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and Its Implications to Second-Order Methods , 2013, SIAM J. Optim..
[17] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[18] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[19] Stephen J. Wright. Coordinate descent algorithms , 2015, Mathematical Programming.
[20] Peter Richtárik,et al. Randomized Iterative Methods for Linear Systems , 2015, SIAM J. Matrix Anal. Appl..
[21] Andrea Montanari,et al. Convergence rates of sub-sampled Newton methods , 2015, NIPS.
[22] Peter Richtárik,et al. Stochastic Dual Ascent for Solving Linear Systems , 2015, ArXiv.
[23] Mark W. Schmidt,et al. Coordinate Descent Converges Faster with the Gauss-Southwell Rule Than Random Selection , 2015, ICML.
[24] Mark W. Schmidt,et al. Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition , 2016, ECML/PKDD.
[25] Peter Richtárik,et al. Coordinate descent with arbitrary sampling I: algorithms and complexity† , 2014, Optim. Methods Softw..
[26] Peter Richtárik,et al. SDNA: Stochastic Dual Newton Ascent for Empirical Risk Minimization , 2015, ICML.
[27] Peng Xu,et al. Sub-sampled Newton Methods with Non-uniform Sampling , 2016, NIPS.
[28] Zeyuan Allen Zhu,et al. Even Faster Accelerated Coordinate Descent Using Non-Uniform Sampling , 2015, ICML.
[29] Peter Richtárik,et al. Coordinate descent with arbitrary sampling II: expected separable overapproximation , 2014, Optim. Methods Softw..
[30] D. Gleich. TRUST REGION METHODS , 2017 .
[31] Robert M. Gower,et al. Randomized Quasi-Newton Updates Are Linearly Convergent Matrix Inversion Algorithms , 2016, SIAM J. Matrix Anal. Appl..
[32] Aurélien Lucchi,et al. Sub-sampled Cubic Regularization for Non-convex Optimization , 2017, ICML.
[33] Yurii Nesterov,et al. Efficiency of the Accelerated Coordinate Descent Method on Structured Optimization Problems , 2017, SIAM J. Optim..
[34] Yurii Nesterov,et al. Regularized Newton Methods for Minimizing Functions with Hölder Continuous Hessians , 2017, SIAM J. Optim..
[35] Martin J. Wainwright,et al. Newton Sketch: A Near Linear-Time Optimization Algorithm with Linear-Quadratic Convergence , 2015, SIAM J. Optim..
[36] Michael I. Jordan,et al. Breaking Locality Accelerates Block Gauss-Seidel , 2017, ICML.
[37] Yurii Nesterov,et al. Lectures on Convex Optimization , 2018 .
[38] Peter Richtárik,et al. SEGA: Variance Reduction via Gradient Sketching , 2018, NeurIPS.
[39] Peter Richtárik,et al. Randomized Block Cubic Newton Method , 2018, ICML.
[40] Martin Jaggi,et al. Global linear convergence of Newton's method without strong-convexity or Lipschitz gradients , 2018, ArXiv.
[41] Robert M. Gower,et al. Accelerated Stochastic Matrix Inversion: General Theory and Speeding up BFGS Rules for Faster Second-Order Optimization , 2018, NeurIPS.
[42] Peter Richtárik,et al. Stochastic Spectral and Conjugate Descent Methods , 2018, NeurIPS.
[43] Michael I. Jordan,et al. Stochastic Cubic Regularization for Fast Nonconvex Optimization , 2017, NeurIPS.
[44] Katya Scheinberg,et al. Global convergence rate analysis of unconstrained optimization methods based on probabilistic models , 2015, Mathematical Programming.
[45] Dmitry Kovalev,et al. RSN: Randomized Subspace Newton , 2019, NeurIPS.
[46] Yurii Nesterov,et al. Accelerated Regularized Newton Methods for Minimizing Composite Convex Functions , 2019, SIAM J. Optim..
[47] Yi Zhou,et al. Sample Complexity of Stochastic Variance-Reduced Cubic Regularization for Nonconvex Optimization , 2018, AISTATS.
[48] Dmitry Kovalev,et al. Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates , 2019, ArXiv.
[49] Doikov Nikita,et al. Inexact basic tensor methods , 2019 .
[50] Convergence Analysis of the Randomized Newton Method with Determinantal Sampling , 2019, ArXiv.
[51] L. Tenorio,et al. Stochastic Subspace Descent , 2019, 1904.01145.
[52] Yair Carmon,et al. Gradient Descent Finds the Cubic-Regularized Nonconvex Newton Step , 2019, SIAM J. Optim..
[53] Martin Jaggi,et al. Efficient Greedy Coordinate Descent for Composite Problems , 2019, AISTATS.
[54] Michael W. Mahoney,et al. Sub-sampled Newton methods , 2018, Math. Program..
[55] Peter Richtárik,et al. Accelerated Coordinate Descent with Arbitrary Sampling and Best Rates for Minibatches , 2018, AISTATS.
[56] A Randomized Coordinate Descent Method with Volume Sampling , 2020, SIAM J. Optim..
[57] Peng Xu,et al. Newton-type methods for non-convex optimization under inexact Hessian information , 2017, Math. Program..
[58] Peter Richtárik,et al. Stochastic Reformulations of Linear Systems: Algorithms and Convergence Theory , 2017, SIAM J. Matrix Anal. Appl..
[59] Yurii Nesterov,et al. Minimizing Uniformly Convex Functions by Cubic Regularization of Newton Method , 2019, Journal of Optimization Theory and Applications.