Trust Region Algorithms and Timestep Selection

Unconstrained optimization problems are closely related to systems of ordinary differential equations (ODEs) with gradient structure. In this work, we prove results that apply to both areas. We analyze the convergence properties of a trust region, or Levenberg--Marquardt, algorithm for optimization. The algorithm may also be regarded as a linearized implicit Euler method with adaptive timestep for gradient ODEs. From the optimization viewpoint, the algorithm is driven directly by the Levenberg--Marquardt parameter rather than the trust region radius. This approach is discussed, for example, in [R. Fletcher, Practical Methods of Optimization, 2nd ed., John Wiley, New York, 1987], but no convergence theory is developed. We give a rigorous error analysis for the algorithm, establishing global convergence and an unusual, extremely rapid, type of superlinear convergence. The precise form of superlinear convergence is exhibited---the ratio of successive displacements from the limit point is bounded above and below by geometrically decreasing sequences. We also show how an inexpensive change to the algorithm leads to quadratic convergence. From the ODE viewpoint, this work contributes to the theory of gradient stability by presenting an algorithm that reproduces the correct global dynamics and gives very rapid local convergence to a stable steady state.

[1]  R. Fletcher Practical Methods of Optimization , 1988 .

[2]  B. Kendall Nonlinear Dynamics and Chaos , 2001 .

[3]  H. C. Yee,et al.  GLOBAL ASYMPTOTIC BEHAVIOR OF ITERATIVE IMPLICIT SCHEMES , 1994 .

[4]  George Hall,et al.  Equilibrium states of Runge Kutta schemes , 1985, TOMS.

[5]  John E. Dennis,et al.  Numerical methods for unconstrained optimization and nonlinear equations , 1983, Prentice Hall series in computational mathematics.

[6]  James M. Ortega,et al.  Iterative solution of nonlinear equations in several variables , 2014, Computer science and applied mathematics.

[7]  Johannes Schropp,et al.  Using dynamical systems methods to solve minimization problems , 1995 .

[8]  Andrew M. Stuart,et al.  The essential stability of local error control for dynamical systems , 1995 .

[9]  Shirley Dex,et al.  JR 旅客販売総合システム(マルス)における運用及び管理について , 1991 .

[10]  J. J. Moré,et al.  A Characterization of Superlinear Convergence and its Application to Quasi-Newton Methods , 1973 .

[11]  M. J. D. Powell,et al.  Nonlinear optimization, 1981 , 1982 .

[12]  U. Helmke,et al.  Optimization and Dynamical Systems , 1994, Proceedings of the IEEE.

[13]  F. Krogh,et al.  Solving Ordinary Differential Equations , 2019, Programming for Computations - Python.

[14]  Desmond J. Higham,et al.  Analysis of the dynamics of local error control via a piecewise continuous residual , 1998 .

[15]  P. Boggs The Solution of Nonlinear Operator Equations by A-stable Integration Techniques , 1970 .

[16]  M. Chu A List of Matrix Flows with Applications , 1994 .

[17]  J. Lambert Numerical Methods for Ordinary Differential Systems: The Initial Value Problem , 1991 .

[18]  Ernst Hairer,et al.  Solving Ordinary Differential Equations I: Nonstiff Problems , 2009 .

[19]  C. Kelley,et al.  Convergence Analysis of Pseudo-Transient Continuation , 1998 .

[20]  Andrew M. Stuart,et al.  Runge-Kutta methods for dissipative and gradient dynamical systems , 1994 .

[21]  S. Goldfeld,et al.  Maximization by Quadratic Hill-Climbing , 1966 .

[22]  Andrew M. Stuart,et al.  Model Problems in Numerical Stability Theory for Initial Value Problems , 1994, SIAM Rev..

[23]  Philip E. Gill,et al.  Practical optimization , 1981 .

[24]  Roger Fletcher,et al.  Practical methods of optimization; (2nd ed.) , 1987 .

[25]  R. Brent,et al.  Fast local convergence with single and multistep methods for nonlinear equations , 1975, The Journal of the Australian Mathematical Society. Series B. Applied Mathematics.

[26]  A. Bloch Hamiltonian and Gradient Flows, Algorithms and Control , 1995 .

[27]  Nicholas J. Higham,et al.  INVERSE PROBLEMS NEWSLETTER , 1991 .