Convergen e Rate of In remental Subgradient Algorithms

We consider a class of subgradient methods for minimizing a convex function that consists of the sum of a large number of component functions. This type of minimization arises in a dual context from Lagrangian relaxation of the coupling constraints of large scale separable problems. The idea is to perform the subgradient iteration incrementally, by sequentially taking steps along the subgradients of the component functions, with intermediate adjustment of the variables after processing each component function. This incremental approach has been very successful in solving large differentiable least squares problems, such as those arising in the training of neural networks, and it has resulted in a much better practical rate of convergence than the steepest descent method.

[1]  Boris Polyak Minimization of unsmooth functionals , 1969 .

[2]  Yuri Ermoliev,et al.  Stochastic Programming Methods , 1976 .

[3]  Jean-Louis Goffin,et al.  The Relaxation Method for Solving Systems of Linear Inequalities , 1980, Math. Oper. Res..

[4]  H. Sherali,et al.  On the choice of step size in subgradient optimization , 1981 .

[5]  Y. Ermoliev Stochastic quasigradient methods and their application to system optimization , 1983 .

[6]  V. F. Dem'yanov,et al.  Nondifferentiable Optimization , 1985 .

[7]  David K. Smith,et al.  Mathematical Programming: Theory and Algorithms , 1986 .

[8]  Jeffery L. Kennington,et al.  A generalization of Polyak's convergence result for subgradient optimization , 1987, Math. Program..

[9]  Yuri Ermoliev,et al.  Numerical techniques for stochastic optimization , 1988 .

[10]  A. Kulikov,et al.  Convex optimization with prescribed accuracy , 1990 .

[11]  Sehun Kim,et al.  Variable target value subgradient method , 1991, Math. Program..

[12]  J. Hiriart-Urruty,et al.  Convex analysis and minimization algorithms , 1993 .

[13]  Claude Lemaréchal,et al.  Convergence of some algorithms for convex minimization , 1993, Math. Program..

[14]  Sehun Kim,et al.  An improved subgradient method for constrained nondifferentiable optimization , 1993, Oper. Res. Lett..

[15]  Dimitri P. Bertsekas,et al.  Nonlinear Programming , 1997 .

[16]  John N. Tsitsiklis,et al.  Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.

[17]  K. Kiwiel The Efficiency of Subgradient Projection Methods for Convex Optimization , 1996 .

[18]  O. Nelles,et al.  An Introduction to Optimization , 1996, IEEE Antennas and Propagation Magazine.

[19]  K. Kiwiel The efficiency of subgradient projection methods for convex optimization, part I: general level methods , 1996 .

[20]  M. Caramanis,et al.  Efficient Lagrangian relaxation algorithms for industry size job-shop scheduling problems , 1998 .

[21]  M. Solodov,et al.  Error Stability Properties of Generalized Gradient-Type Algorithms , 1998 .

[22]  Torbjörn Larsson,et al.  The Efficiency of Ballstep Subgradient Level Methods for Convex Optimization , 1999, Math. Oper. Res..

[23]  Jean-Louis Goffin,et al.  Convergence of a simple subgradient level method , 1999, Math. Program..

[24]  X. Zhao,et al.  Surrogate Gradient Algorithm for Lagrangian Relaxation , 1999 .

[25]  Dimitri P. Bertsekas,et al.  Incremental Subgradient Methods for Nondifferentiable Optimization , 2001, SIAM J. Optim..