On the Goldstein-Levitin-Polyak gradient projection method

This paper considers some aspects of a gradient projection method proposed by Goldstein [1], Levitin and Polyak [3] and more recently, in a less general context, by Mc-Cormick [10]. We propose and analyze some convergent stepsize rules to be used in conjunction with the method. These rules are similar in spirit with the efficient Armijo rule for the method of steepest descent and under mild assumptions they have the desirable property that they identify the set of active inequality constraints in a finite number of iterations. As a result the method may be converted towards the end of the process to a conjugate direction, Quasi-Newton or Newton's method and achieve the attendant superlinear convergence rate. As an example we propose a quadratically convergent combination of the method with Newton's method. Such combined methods appear to be very efficient for large scale problems with many simple constraints such as those often appearing in optimal control.