Reduced-Hessian Quasi-Newton Methods for Unconstrained Optimization

Quasi-Newton methods are reliable and efficient on a wide range of problems, but they can require many iterations if the problem is ill-conditioned or if a poor initial estimate of the Hessian is used. In this paper, we discuss methods designed to be more efficient in these situations. All the methods to be considered exploit the fact that quasi-Newton methods accumulate approximate second-derivative information in a sequence of expanding subspaces. Associated with each of these subspaces is a certain reduced approximate Hessian that provides a complete but compact representation of the second derivative information approximated up to that point. Algorithms that compute an explicit reduced-Hessian approximation have two important advantages over conventional quasi-Newton methods. First, the amount of computation for each iteration is significantly less during the early stages. This advantage is increased by forcing the iterates to linger on a manifold whose dimension can be significantly smaller than the subspace in which curvature has been accumulated. Second, approximate curvature along directions that lie off the manifold can be reinitialized as the iterations proceed, thereby reducing the influence of a poor initial estimate of the Hessian. These advantages are illustrated by extensive numerical results from problems in the CUTE test set. Our experiments provide strong evidence that reduced-Hessian quasi-Newton methods are more efficient and robust than conventional BFGS methods and some recently proposed extensions.

[1]  D. F. Shanno,et al.  Matrix conditioning and nonlinear optimization , 1978, Math. Program..

[2]  Dirk Siegel,et al.  Modifying the BFGS update by a new column scaling technique , 1994, Math. Program..

[3]  Nicholas I. M. Gould,et al.  CUTE: constrained and unconstrained testing environment , 1995, TOMS.

[4]  Jorge Nocedal,et al.  On the limited memory BFGS method for large scale optimization , 1989, Math. Program..

[5]  Roger Fletcher,et al.  A Rapidly Convergent Descent Method for Minimization , 1963, Comput. J..

[6]  D. Goldfarb Factorized variable metric methods for unconstrained optimization , 1976 .

[7]  Jorge Nocedal,et al.  Automatic Column Scaling Strategies for Quasi-Newton Methods , 1993, SIAM J. Optim..

[8]  G. Golub,et al.  Iterative solution of linear systems , 1991, Acta Numerica.

[9]  C. Kelley Iterative Methods for Linear and Nonlinear Equations , 1987 .

[10]  Gene H. Golub,et al.  Matrix computations , 1983 .

[11]  Shirley Dex,et al.  JR 旅客販売総合システム(マルス)における運用及び管理について , 1991 .

[12]  D. Luenberger,et al.  Self-Scaling Variable Metric (SSVM) Algorithms , 1974 .

[13]  John L. Nazareth The method of successive affine reduction for nonlinear minimization , 1986, Math. Program..

[14]  Philip E. Gill,et al.  Practical optimization , 1981 .

[15]  M. J. D. Powell,et al.  Updating conjugate directions by the BFGS formula , 1987, Math. Program..

[16]  Roger Fletcher,et al.  Practical methods of optimization; (2nd ed.) , 1987 .

[17]  David J. Thuente,et al.  Line search algorithms with guaranteed sufficient decrease , 1994, TOMS.

[18]  Jorge J. Moré,et al.  Testing Unconstrained Optimization Software , 1981, TOMS.

[19]  J. J. Moré,et al.  Newton's Method , 1982 .

[20]  R. Fletcher Practical Methods of Optimization , 1988 .

[21]  L. Nazareth A Relationship between the BFGS and Conjugate Gradient Algorithms and Its Implications for New Algorithms , 1979 .

[22]  J. Nocedal Updating Quasi-Newton Matrices With Limited Storage , 1980 .

[23]  R. Schnabel,et al.  A NEW DERIVATION OF SYMMETRIC POSITIVE DEFINITE SECANT UPDATES , 1980 .

[24]  G. Stewart,et al.  Reorthogonalization and stable algorithms for updating the Gram-Schmidt QR factorization , 1976 .

[25]  Michael A. Saunders,et al.  User''s guide for NPSOL (Ver-sion 4.0): A FORTRAN package for nonlinear programming , 1984 .