Preconditioning and globalizing conjugate gradients in dual space for quadratically penalized nonlinear-least squares problems

When solving nonlinear least-squares problems, it is often useful to regularize the problem using a quadratic term, a practice which is especially common in applications arising in inverse calculations. A solution method derived from a trust-region Gauss-Newton algorithm is analyzed for such applications, where, contrary to the standard algorithm, the least-squares subproblem solved at each iteration of the method is rewritten as a quadratic minimization subject to linear equality constraints. This allows the exploitation of duality properties of the associated linearized problems. This paper considers a recent conjugate-gradient-like method which performs the quadratic minimization in the dual space and produces, in exact arithmetic, the same iterates as those produced by a standard conjugate-gradients method in the primal space. This dual algorithm is computationally interesting whenever the dimension of the dual space is significantly smaller than that of the primal space, yielding gains in terms of both memory usage and computational cost. The relation between this dual space solver and PSAS (Physical-space Statistical Analysis System), another well-known dual space technique used in data assimilation problems, is explained. The use of an effective preconditioning technique is proposed and refined convergence bounds derived, which results in a practical solution method. Finally, stopping rules adequate for a trust-region solver are proposed in the dual space, providing iterates that are equivalent to those obtained with a Steihaug-Toint truncated conjugate-gradient method in the primal space.

[1]  Jérôme Vialard,et al.  Three- and Four-Dimensional Variational Assimilation with a General Circulation Model of the Tropical Pacific Ocean. Part I: Formulation, Internal Diagnostics, and Consistency Checks , 2003 .

[2]  Stephen J. Wright,et al.  Numerical Optimization , 2018, Fundamental Statistical Inference.

[3]  A. Tarantola Inverse problem theory : methods for data fitting and model parameter estimation , 1987 .

[4]  Nicholas I. M. Gould,et al.  Trust Region Methods , 2000, MOS-SIAM Series on Optimization.

[5]  Martin Stoll,et al.  Combination preconditioning and self-adjointness in non-standard inner products with application to saddle point problems , 2007 .

[6]  P. Toint,et al.  Adaptive cubic overestimation methods for unconstrained optimization. Part I: motivation, convergence and numerical results , 2008 .

[7]  Stephen J. Wright,et al.  Data assimilation in weather forecasting: a case study in PDE-constrained optimization , 2009 .

[8]  Gene H. Golub,et al.  Matrix computations , 1983 .

[9]  François-Xavier Roux,et al.  Acceleration of the outer conjugate gradient by reorthogonalization for a domain decomposition method for structural analysis problems , 1989, ICS '89.

[10]  M. Arioli,et al.  A stopping criterion for the conjugate gradient algorithm in a finite element method framework , 2000, Numerische Mathematik.

[11]  Samuel Buis,et al.  Intercomparison of the primal and dual formulations of variational data assimilation , 2008 .

[12]  Thomas Kaminski,et al.  Recipes for adjoint code construction , 1998, TOMS.

[13]  Martin Stoll,et al.  Combination Preconditioning and the Bramble-Pasciak+ Preconditioner , 2008, SIAM J. Matrix Anal. Appl..

[14]  A. Morelli Inverse Problem Theory , 2010 .

[15]  P. Courtier,et al.  A strategy for operational implementation of 4D‐Var, using an incremental approach , 1994 .

[16]  Barbara Kaltenbacher,et al.  Iterative Solution Methods , 2015, Handbook of Mathematical Methods in Imaging.

[17]  Carl Tim Kelley,et al.  Iterative methods for optimization , 1999, Frontiers in applied mathematics.

[18]  Serge Gratton,et al.  An observation‐space formulation of variational assimilation using a restricted preconditioned conjugate gradient algorithm , 2009 .

[19]  Jorge Nocedal,et al.  Automatic Preconditioning by Limited Memory Quasi-Newton Updating , 1999, SIAM J. Optim..

[20]  Paul E. Saylor,et al.  The Role of the Inner Product in Stopping Criteria for Conjugate Gradient Iterations , 2001 .

[21]  John E. Dennis,et al.  Numerical methods for unconstrained optimization and nonlinear equations , 1983, Prentice Hall series in computational mathematics.

[22]  Åke Björck,et al.  Numerical methods for least square problems , 1996 .

[23]  Serge Gratton,et al.  Limited‐memory preconditioners, with application to incremental four‐dimensional variational data assimilation , 2008 .

[24]  Danny C. Sorensen,et al.  Accelerating the LSTRS Algorithm , 2010, SIAM J. Sci. Comput..

[25]  Serge Gratton,et al.  Approximate Gauss-Newton Methods for Nonlinear Least Squares Problems , 2007, SIAM J. Optim..

[26]  P. Toint,et al.  Adaptive cubic overestimation methods for unconstrained optimization , 2007 .

[27]  Nicholas I. M. Gould,et al.  Adaptive cubic regularisation methods for unconstrained optimization. Part I: motivation, convergence and numerical results , 2011, Math. Program..

[28]  Danny C. Sorensen,et al.  Algorithm 873: LSTRS: MATLAB software for large-scale trust-region subproblems and regularization , 2008, TOMS.

[29]  Philippe Courtier,et al.  Dual formulation of four‐dimensional variational assimilation , 1997 .

[30]  Stephen J. Wright,et al.  Numerical Optimization (Springer Series in Operations Research and Financial Engineering) , 2000 .