Jacobian-Free Implicit Inner-Iteration Preconditioner for Nonlinear Least Squares Problems

Nonlinear least squares (NLS) problems arise in many applications. The common solvers require to compute and store the corresponding Jacobian matrix explicitly, which is too expensive for large problems. Recently, some Jacobian-free (or matrix free) methods were proposed, but most of these methods are not really Jacobian free since the full or partial Jacobian matrix still needs to be computed in some iteration steps. In this paper, we propose an effective real Jacobian free method especially for large NLS problems, which is realized by the novel combination of using automatic differentiation for $$J(\mathbf{x})\mathbf{v}$$J(x)v and $$J(\mathbf{x})^T\mathbf{v}$$J(x)Tv along with the implicit iterative preconditioning ideas. Together, they yield a new and effective three-level iterative approach. In the outer level, the dogleg/trust region method is employed to solve the NLS problem. At each iteration of the dogleg method, we adopt the iterative linear least squares (LLS) solvers, CGLS or BA-GMRES method, to solve the LLS problem generated at each step of the dogleg method as the middle iteration. In order to accelerate the convergence of the iterative LLS solver, we propose an implicit inner iteration preconditioner based on the weighted Jacobi method. Compared to the existing Jacobian-free methods, our proposed three-level method need not compute any part of the Jacobian matrix explicitly in any iteration step. Furthermore, our method does not rely on the sparsity or structure pattern of the Jacobian, gradient or Hessian matrix. In other words, our method also works well for dense Jacobian matrices. Numerical experiments show the superiority of our proposed method.

[1]  Wei Xu,et al.  A secant method for nonlinear least-squares minimization , 2012, Comput. Optim. Appl..

[2]  Miroslav Tuma,et al.  Efficient Preconditioning of Sequences of Nonsymmetric Linear Systems , 2007, SIAM J. Sci. Comput..

[3]  Daniela di Serafino,et al.  Efficient Preconditioner Updates for Shifted Linear Systems , 2011, SIAM J. Sci. Comput..

[4]  T. Coleman,et al.  The Efficient Application of Automatic Differentiation for Computing Gradients in Financial Applications , 2016 .

[5]  Miroslav Tuma,et al.  Preconditioner updates for solving sequences of linear systems in matrix‐free environment , 2010, Numer. Linear Algebra Appl..

[6]  M. Hestenes,et al.  Methods of conjugate gradients for solving linear systems , 1952 .

[7]  Daniele Bertaccini,et al.  Nonsymmetric Preconditioner Updates in Newton-Krylov Methods for Nonlinear Systems , 2011, SIAM J. Sci. Comput..

[8]  Homer F. Walker,et al.  Inexact Newton Dogleg Methods , 2005, SIAM J. Numer. Anal..

[9]  Miroslav Tuma,et al.  Preconditioned Iterative Methods for Solving Linear Least Squares Problems , 2014, SIAM J. Sci. Comput..

[10]  Gene H. Golub,et al.  Matrix computations (3rd ed.) , 1996 .

[11]  Wei Xu,et al.  Fast (Structured) Newton Computations , 2008, SIAM J. Sci. Comput..

[12]  L. Dixon,et al.  Truncated Newton method for sparse unconstrained optimization using automatic differentiation , 1989 .

[13]  Andreas Griewank,et al.  Evaluating derivatives - principles and techniques of algorithmic differentiation, Second Edition , 2000, Frontiers in applied mathematics.

[14]  A. Sluis Condition numbers and equilibration of matrices , 1969 .

[15]  S. Nash Preconditioning of Truncated-Newton Methods , 1985 .

[16]  Tetsuya Sakurai,et al.  An Auto-Tuning Technique of the Weighted Jacobi-Type Iteration Used for Preconditioners of Krylov Subspace Methods , 2012, 2012 IEEE 6th International Symposium on Embedded Multicore SoCs.

[17]  Yousef Saad,et al.  Iterative methods for sparse linear systems , 2003 .

[18]  Jorge Nocedal,et al.  Automatic Preconditioning by Limited Memory Quasi-Newton Updating , 1999, SIAM J. Optim..

[19]  John E. Dennis,et al.  Numerical methods for unconstrained optimization and nonlinear equations , 1983, Prentice Hall series in computational mathematics.

[20]  Åke Björck,et al.  Numerical methods for least square problems , 1996 .

[21]  Kaj Madsen,et al.  Introduction to Optimization and Data Fitting , 2008 .

[22]  Keiichi Morikuni,et al.  Inner-Iteration Krylov Subspace Methods for Least Squares Problems , 2013, SIAM J. Matrix Anal. Appl..

[23]  Jacek Gondzio,et al.  A Matrix-Free Preconditioner for Sparse Symmetric Positive Definite Systems and Least-Squares Problems , 2013, SIAM J. Sci. Comput..

[24]  Gene H. Golub,et al.  Matrix computations , 1983 .

[25]  Jun-Feng Yin,et al.  GMRES Methods for Least Squares Problems , 2010, SIAM J. Matrix Anal. Appl..

[26]  Ying Chen,et al.  A Jacobian-free Newton-GMRES(m) method with adaptive preconditioner and its application for power flow calculations , 2006, IEEE Transactions on Power Systems.

[27]  T. Coleman,et al.  Reconstructing the Unknown Local Volatility Function , 1999 .

[28]  D. Keyes,et al.  Jacobian-free Newton-Krylov methods: a survey of approaches and applications , 2004 .

[29]  W. Xu,et al.  Efficient Preconditioners for Newton-GMRES Method With Application to Power Flow Study , 2013, IEEE Transactions on Power Systems.

[30]  Martin B. van Gijzen,et al.  IDR(s): A Family of Simple and Fast Algorithms for Solving Large Nonsymmetric Systems of Linear Equations , 2008, SIAM J. Sci. Comput..

[31]  Y. Saad,et al.  GMRES: a generalized minimal residual algorithm for solving nonsymmetric linear systems , 1986 .

[32]  Keiichi Morikuni,et al.  Convergence of Inner-Iteration GMRES Methods for Rank-Deficient Least Squares Problems , 2015, SIAM J. Matrix Anal. Appl..

[33]  Wei Xu,et al.  Efficient (Partial) Determination of Derivative Matrices via Automatic Differentiation , 2013, SIAM J. Sci. Comput..

[34]  Wei Xu,et al.  Solving nonlinear equations with the Newton–Krylov method based on automatic differentiation , 2014, Optim. Methods Softw..