WORST-CASE EVALUATION COMPLEXITY AND OPTIMALITY OF SECOND-ORDER METHODS FOR NONCONVEX SMOOTH OPTIMIZATION

We establish or refute the optimality of inexact second-order methods for unconstrained nonconvex optimization from the point of view of worst-case evaluation complexity, improving and generalizing the results of Cartis, Gould and Toint (2010,2011). To this aim, we consider a new general class of inexact second-order algorithms for unconstrained optimization that includes regularization and trust-region variations of Newton's method as well as of their linesearch variants. For each method in this class and arbitrary accuracy threshold $\epsilon \in (0,1)$, we exhibit a smooth objective function with bounded range, whose gradient is globally Lipschitz continuous and whose Hessian is $\alpha-$H\"older continuous (for given $\alpha\in [0,1]$), for which the method in question takes at least $\lfloor\epsilon^{-(2+\alpha)/(1+\alpha)}\rfloor$ function evaluations to generate a first iterate whose gradient is smaller than $\epsilon$ in norm. Moreover, we also construct another function on which Newton's takes $\lfloor\epsilon^{-2}\rfloor$ evaluations, but whose Hessian is Lipschitz continuous on the path of iterates. These examples provide lower bounds on the worst-case evaluation complexity of methods in our class when applied to smooth problems satisfying the relevant assumptions. Furthermore, for $\alpha=1$, this lower bound is of the same order in $\epsilon$ as the upper bound on the worst-case evaluation complexity of the cubic and other methods in a class of methods proposed in Curtis, Robinson and samadi (2017) or in Royer and Wright (2017), thus implying that these methods have optimal worst-case evaluation complexity within a wider class of second-order methods, and that Newton's method is suboptimal.

[1]  S. Goldfeld,et al.  Maximization by Quadratic Hill-Climbing , 1966 .

[2]  John E. Dennis,et al.  Numerical methods for unconstrained optimization and nonlinear equations , 1983, Prentice Hall series in computational mathematics.

[3]  Stephen A. Vavasis,et al.  Black-Box Complexity of Local Minimization , 1993, SIAM J. Optim..

[4]  S. Lucidi,et al.  A Linesearch Algorithm with Memory for Unconstrained Optimization , 1998 .

[5]  Stephen J. Wright,et al.  Numerical Optimization , 2018, Fundamental Statistical Inference.

[6]  Stephen J. Wright,et al.  Numerical Optimization (Springer Series in Operations Research and Financial Engineering) , 2000 .

[7]  Yurii Nesterov,et al.  Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.

[8]  Yurii Nesterov,et al.  Cubic regularization of Newton method and its global performance , 2006, Math. Program..

[9]  Peter Deuflhard,et al.  Affine conjugate adaptive Newton methods for nonlinear elastomechanics , 2007, Optim. Methods Softw..

[10]  P. Toint,et al.  Adaptive cubic overestimation methods for unconstrained optimization , 2007 .

[11]  Serge Gratton,et al.  Recursive Trust-Region Methods for Multiscale Nonlinear Optimization , 2008, SIAM J. Optim..

[12]  P. Toint,et al.  Adaptive cubic overestimation methods for unconstrained optimization. Part I: motivation, convergence and numerical results , 2008 .

[13]  Nobuo Yamashita,et al.  On a Global Complexity Bound of the Levenberg-Marquardt Method , 2010, J. Optim. Theory Appl..

[14]  Nicholas I. M. Gould,et al.  On the Complexity of Steepest Descent, Newton's and Regularized Newton's Methods for Nonconvex Unconstrained Optimization Problems , 2010, SIAM J. Optim..

[15]  N. Yamashita,et al.  Convergence Properties of the Regularized Newton Method for the Unconstrained Nonconvex Optimization , 2010 .

[16]  Yinyu Ye,et al.  A note on the complexity of Lp minimization , 2011, Math. Program..

[17]  Nicholas I. M. Gould,et al.  Adaptive cubic regularisation methods for unconstrained optimization. Part I: motivation, convergence and numerical results , 2011, Math. Program..

[18]  Nicholas I. M. Gould,et al.  On the Evaluation Complexity of Composite Function Minimization with Applications to Nonconvex Nonlinear Programming , 2011, SIAM J. Optim..

[19]  Nicholas I. M. Gould,et al.  Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity , 2011, Math. Program..

[20]  P. Toint,et al.  An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity , 2012 .

[21]  Nicholas I. M. Gould,et al.  Updating the regularization parameter in the adaptive cubic regularization algorithm , 2012, Comput. Optim. Appl..

[22]  Lue Li,et al.  A trust region algorithm with adaptive cubic regularization methods for nonsmooth convex minimization , 2012, Comput. Optim. Appl..

[23]  Nicholas I. M. Gould,et al.  Complexity bounds for second-order optimality in unconstrained optimization , 2012, J. Complex..

[24]  Ph. L. Toint,et al.  On the complexity of the steepest-descent with exact linesearches , 2012 .

[25]  Nicholas I. M. Gould,et al.  On the Evaluation Complexity of Cubic Regularization Methods for Potentially Rank-Deficient Nonlinear Least-Squares Problems and Its Relevance to Constrained Nonlinear Optimization , 2013, SIAM J. Optim..

[26]  Katya Scheinberg,et al.  Complexity of Inexact Proximal Newton methods , 2013, ArXiv.

[27]  Luís Nunes Vicente,et al.  Worst case complexity of direct search , 2013, EURO J. Comput. Optim..

[28]  Xiaojun Chen,et al.  Worst-Case Complexity of Smoothing Quadratic Regularization Methods for Non-Lipschitzian Optimization , 2013, SIAM J. Optim..

[29]  Florian Jarre,et al.  On Nesterov's smooth Chebyshev–Rosenbrock function , 2013, Optim. Methods Softw..

[30]  Nicholas I. M. Gould,et al.  On the complexity of finding first-order critical points in constrained nonlinear optimization , 2014, Math. Program..

[31]  Xiaojun Chen,et al.  Complexity analysis of interior point algorithms for non-Lipschitz and nonconvex minimization , 2015, Math. Program..

[32]  Serge Gratton,et al.  Direct Search Based on Probabilistic Descent , 2015, SIAM J. Optim..

[33]  Ya-xiang Yuan,et al.  On the convergence and worst-case complexity of trust-region and regularization methods for unconstrained optimization , 2015, Math. Program..

[34]  Nicholas I. M. Gould,et al.  On the Evaluation Complexity of Constrained Nonlinear Least-Squares and General Constrained Nonlinear Optimization Using Second-Order Methods , 2015, SIAM J. Numer. Anal..

[35]  J. Dussault Simple unified convergence proofs for Trust Region and a new ARC variant , 2015 .

[36]  Jean-Pierre Dussault,et al.  Scalable adaptative cubic regularization methods , 2015, 2103.16659.

[37]  Xiaojun Chen,et al.  NON-LIPSCHITZ OPTIMIZATION FOR IMAGE RESTORATION , 2015 .

[38]  P. Toint,et al.  Improved worst-case evaluation complexity for potentially rank-deficient nonlinear least-Euclidean-norm problems using higher-order regularized models , 2015 .

[39]  Marco Sciandrone,et al.  On the use of iterative methods in cubic regularization for unconstrained optimization , 2015, Comput. Optim. Appl..

[40]  P. Toint,et al.  Worst-case evaluation complexity of non-monotone gradient-related algorithms for unconstrained optimization , 2015 .

[41]  Jin Yun Yuan,et al.  Nonlinear Stepsize Control Algorithms: Complexity Bounds for First- and Second-Order Optimality , 2016, J. Optim. Theory Appl..

[42]  Saeed Ghadimi,et al.  Accelerated gradient methods for nonconvex nonlinear and stochastic programming , 2013, Mathematical Programming.

[43]  Luís N. Vicente,et al.  On the optimal order of worst case complexity of direct search , 2016, Optim. Lett..

[44]  Luís Nunes Vicente,et al.  Trust-Region Methods Without Using Derivatives: Worst Case Complexity and the NonSmooth Case , 2016, SIAM J. Optim..

[45]  Anima Anandkumar,et al.  Efficient approaches for escaping higher order saddle points in non-convex optimization , 2016, COLT.

[46]  Yair Carmon,et al.  Gradient Descent Efficiently Finds the Cubic-Regularized Non-Convex Newton Step , 2016, ArXiv.

[47]  Mingyi Hong,et al.  Decomposing Linearly Constrained Nonconvex Problems by a Proximal Primal Dual Approach: Algorithms, Convergence, and Applications , 2016, ArXiv.

[48]  Katya Scheinberg,et al.  Practical inexact proximal quasi-Newton method with global complexity analysis , 2013, Mathematical Programming.

[49]  Tengyu Ma,et al.  Finding Approximate Local Minima for Nonconvex Optimization in Linear Time , 2016, ArXiv.

[50]  Y. Nesterov,et al.  Globally Convergent Second-order Schemes for Minimizing Twice-differentiable Functions , 2016 .

[51]  Marco Sciandrone,et al.  A cubic regularization algorithm for unconstrained optimization using line search and nonmonotone techniques , 2016, Optim. Methods Softw..

[52]  José Mario Martínez,et al.  Evaluation Complexity for Nonlinear Constrained Optimization Using Unscaled KKT Conditions and High-Order Models , 2016, SIAM J. Optim..

[53]  D. Gleich TRUST REGION METHODS , 2017 .

[54]  José Mario Martínez,et al.  Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models , 2017, Math. Program..

[55]  Daniel P. Robinson,et al.  An Inexact Regularized Newton Framework with a Worst-Case Iteration Complexity of $\mathcal{O}(\epsilon^{-3/2})$ for Nonconvex Optimization , 2017, 1708.00475.

[56]  Daniel P. Robinson,et al.  A trust region algorithm with a worst-case iteration complexity of O(ϵ-3/2)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{docume , 2016, Mathematical Programming.

[57]  Yair Carmon,et al.  "Convex Until Proven Guilty": Dimension-Free Acceleration of Gradient Descent on Non-Convex Functions , 2017, ICML.

[58]  José Mario Martínez,et al.  On High-order Model Regularization for Constrained Optimization , 2017, SIAM J. Optim..

[59]  Serge Gratton,et al.  On the use of the energy norm in trust-region and adaptive cubic regularization subproblems , 2017, Comput. Optim. Appl..

[60]  Nicholas I. M. Gould,et al.  Improved second-order evaluation complexity for unconstrained nonlinear optimization using high-order regularized models , 2017, ArXiv.

[61]  Hong Wang,et al.  Partially separable convexly-constrained optimization with non-Lipschitzian singularities and its complexity , 2017, ArXiv.

[62]  José Mario Martínez,et al.  Cubic-regularization counterpart of a variable-norm trust-region method for unconstrained minimization , 2017, J. Glob. Optim..

[63]  Tengyu Ma,et al.  Finding approximate local minima faster than gradient descent , 2016, STOC.

[64]  José Mario Martínez,et al.  On Regularization and Active-set Methods with Complexity for Constrained Optimization , 2018, SIAM J. Optim..

[65]  Stephen J. Wright,et al.  Complexity Analysis of Second-Order Line-Search Algorithms for Smooth Nonconvex Optimization , 2017, SIAM J. Optim..

[66]  Daniel P. Robinson,et al.  Complexity Analysis of a Trust Funnel Algorithm for Equality Constrained Optimization , 2017, SIAM J. Optim..

[67]  Frank E. Curtis An inexact regularized Newton framework with a worst-case iteration complexity of O(ε−3/2) for nonconvex optimization , 2018 .

[68]  Nicholas I. M. Gould,et al.  Second-Order Optimality and Beyond: Characterization and Evaluation Complexity in Convexly Constrained Nonlinear Optimization , 2018, Found. Comput. Math..

[69]  Katya Scheinberg,et al.  Global convergence rate analysis of unconstrained optimization methods based on probabilistic models , 2015, Mathematical Programming.

[70]  Nicholas I. M. Gould,et al.  Universal regularization methods - varying the power, the smoothness and the accuracy , 2018, 1811.07057.

[71]  Shiqian Ma,et al.  Structured nonconvex and nonsmooth optimization: algorithms and iteration complexity analysis , 2016, Computational Optimization and Applications.

[72]  Nicholas I. M. Gould,et al.  Optimality of orders one to three and beyond: characterization and evaluation complexity in constrained nonconvex optimization , 2017, J. Complex..

[73]  P. Absil,et al.  Erratum to: ``Global rates of convergence for nonconvex optimization on manifolds'' , 2016, IMA Journal of Numerical Analysis.

[74]  A concise second-order evaluation complexity for unconstrained nonlinear optimization using high-order regularized models , 2019 .

[75]  Serge Gratton,et al.  A decoupled first/second-order steps technique for nonconvex nonlinear unconstrained optimization with improved complexity bounds , 2020, Math. Program..

[76]  Peng Xu,et al.  Newton-type methods for non-convex optimization under inexact Hessian information , 2017, Math. Program..

[77]  Francisco Facchinei,et al.  Ghost Penalties in Nonconvex Constrained Optimization: Diminishing Stepsizes and Iteration Complexity , 2017, Math. Oper. Res..