Convergence Rate Analysis of a Sequential Convex Programming Method with Line Search for a Class of Constrained Difference-of-Convex Optimization Problems

In this paper, we study the sequential convex programming method with monotone line search (SCP$_{ls}$) in [34] for a class of difference-of-convex (DC) optimization problems with multiple smooth inequality constraints. The SCP$_{ls}$ is a representative variant of moving-ball-approximation-type algorithms [4,8,11,39] for constrained optimization problems. We analyze the convergence rate of the sequence generated by SCP$_{ls}$ in both nonconvex and convex settings by imposing suitable Kurdyka-Lojasiewicz (KL) assumptions. Specifically, in the nonconvex settings, we assume that a special potential function related to the objective and the constraints is a KL function, while in the convex settings we impose KL assumptions directly on the extended objective function (i.e., sum of the objective and the indicator function of the constraint set). A relationship between these two different KL assumptions is established in the convex settings under additional differentiability assumptions. We also discuss how to deduce the KL exponent of the extended objective function from its Lagrangian in the convex settings, under additional assumptions on the constraint functions. Thanks to this result, the extended objectives of some constrained optimization models such as minimizing $\ell_1$ subject to logistic/Poisson loss are found to be KL functions with exponent $\frac12$ under mild assumptions. To illustrate how our results can be applied, we consider SCP$_{ls}$ for minimizing $\ell_{1-2}$ [43] subject to residual error measured by $\ell_2$ norm/Lorentzian norm [19]. We first discuss how the various conditions required in our analysis can be verified, and then perform numerical experiments to illustrate the convergence behaviors of SCP$_{ls}$.

[1]  E. Candès The restricted isometry property and its implications for compressed sensing , 2008 .

[2]  Peter Ochs,et al.  Unifying Abstract Inexact Convergence Theorems and Block Coordinate Variable Metric iPiano , 2016, SIAM J. Optim..

[3]  G. Zou,et al.  A modified poisson regression approach to prospective studies with binary data. , 2004, American journal of epidemiology.

[4]  Bruce W. Suter,et al.  From error bounds to the complexity of first-order descent methods for convex functions , 2015, Math. Program..

[5]  Jack Xin,et al.  Minimization of ℓ1-2 for Compressed Sensing , 2015, SIAM J. Sci. Comput..

[6]  Olvi L. Mangasarian,et al.  Nonlinear Programming , 1969 .

[7]  Yi Yang,et al.  Sequential Convex Approximations to Joint Chance Constrained Programs: A Monte Carlo Approach , 2011, Oper. Res..

[8]  Zhaosong Lu,et al.  Sequential Convex Programming Methods for A Class of Structured Nonlinear Programming , 2012, ArXiv.

[9]  F. Clarke Optimization And Nonsmooth Analysis , 1983 .

[10]  E. Frome The analysis of rates using Poisson regression models. , 1983, Biometrics.

[11]  F. Facchinei,et al.  Finite-Dimensional Variational Inequalities and Complementarity Problems , 2003 .

[12]  D. Wooff Logistic Regression: a Self-learning Text, 2nd edn , 2004 .

[13]  K. Kurdyka On gradients of functions definable in o-minimal structures , 1998 .

[14]  Edouard Pauwels,et al.  Majorization-Minimization Procedures and Convergence of SQP Methods for Semi-Algebraic and Tame Programs , 2014, Math. Oper. Res..

[15]  Benar Fux Svaiter,et al.  Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward–backward splitting, and regularized Gauss–Seidel methods , 2013, Math. Program..

[16]  José Mario Martínez,et al.  Nonmonotone Spectral Projected Gradient Methods on Convex Sets , 1999, SIAM J. Optim..

[17]  Adrian S. Lewis,et al.  The [barred L]ojasiewicz Inequality for Nonsmooth Subanalytic Functions with Applications to Subgradient Dynamical Systems , 2006, SIAM J. Optim..

[18]  Hédy Attouch,et al.  On the convergence of the proximal algorithm for nonsmooth functions involving analytic features , 2008, Math. Program..

[19]  丸山 徹 Convex Analysisの二,三の進展について , 1977 .

[20]  Bo Wen,et al.  A proximal difference-of-convex algorithm with extrapolation , 2016, Computational Optimization and Applications.

[21]  Jieping Ye,et al.  A General Iterative Shrinkage and Thresholding Algorithm for Non-convex Regularized Optimization Problems , 2013, ICML.

[22]  Marc Teboulle,et al.  A dual method for minimizing a nonsmooth objective over one smooth inequality constraint , 2016, Math. Program..

[23]  Marc Teboulle,et al.  A Moving Balls Approximation Method for a Class of Smooth Constrained Minimization Problems , 2010, SIAM J. Optim..

[24]  David W. Hosmer,et al.  Applied Logistic Regression , 1991 .

[25]  Kenneth E. Barner,et al.  Robust Sampling and Reconstruction Methods for Sparse Signals in the Presence of Impulsive Noise , 2010, IEEE Journal of Selected Topics in Signal Processing.

[26]  Marc Teboulle,et al.  Proximal alternating linearized minimization for nonconvex and nonsmooth problems , 2013, Mathematical Programming.

[27]  Chong Li,et al.  The SECQ, Linear Regularity, and the Strong CHIP for an Infinite System of Closed Convex Sets in Normed Linear Spaces , 2007, SIAM J. Optim..

[28]  A. Iusem On the convergence properties of the projected gradient method for convex optimization , 2003 .

[29]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[30]  Jonathan M. Borwein,et al.  Analysis of the Convergence Rate for the Cyclic Projection Algorithm Applied to Basic Semialgebraic Convex Sets , 2013, SIAM J. Optim..

[31]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[32]  Michael P. Friedlander,et al.  Probing the Pareto Frontier for Basis Pursuit Solutions , 2008, SIAM J. Sci. Comput..

[33]  S. Landau Book Review: Logistic regression: a self-learning text, 2nd edition , 2004 .

[34]  Brian M. Sadler,et al.  Robust compressive sensing of sparse signals: a review , 2016, EURASIP J. Adv. Signal Process..

[35]  J. Borwein,et al.  Techniques of variational analysis , 2005 .

[36]  Hédy Attouch,et al.  Proximal Alternating Minimization and Projection Methods for Nonconvex Problems: An Approach Based on the Kurdyka-Lojasiewicz Inequality , 2008, Math. Oper. Res..

[37]  Zheng Chen,et al.  The multiproximal linearization method for convex composite problems , 2017, Math. Program..

[38]  Ting Kei Pong,et al.  Deducing Kurdyka-{\L}ojasiewicz exponent via inf-projection , 2019, 1902.03635.

[39]  Guoyin Li,et al.  Calculus of the Exponent of Kurdyka–Łojasiewicz Inequality and Its Applications to Linear Convergence of First-Order Methods , 2016, Foundations of Computational Mathematics.

[40]  Adrian S. Lewis,et al.  Convex Analysis And Nonlinear Optimization , 2000 .

[41]  Guoyin Li,et al.  Douglas–Rachford splitting for nonconvex optimization with application to nonconvex feasibility problems , 2014, Math. Program..

[42]  J. Borwein,et al.  Two-Point Step Size Gradient Methods , 1988 .

[43]  Diane Lambert,et al.  Zero-inflacted Poisson regression, with an application to defects in manufacturing , 1992 .