Feasible Direction Interior-Point Technique for Nonlinear Optimization

We propose a feasible direction approach for the minimization by interior-point algorithms of a smooth function under smooth equality and inequality constraints. It consists of the iterative solution in the primal and dual variables of the Karush–Kuhn–Tucker first-order optimality conditions. At each iteration, a descent direction is defined by solving a linear system. In a second stage, the linear system is perturbed so as to deflect the descent direction and obtain a feasible descent direction. A line search is then performed to get a new interior point and ensure global convergence. Based on this approach, first-order, Newton, and quasi-Newton algorithms can be obtained. To introduce the method, we consider first the inequality constrained problem and present a globally convergent basic algorithm. Particular first-order and quasi-Newton versions of this algorithm are also stated. Then, equality constraints are included. This method, which is simple to code, does not require the solution of quadratic programs and it is neither a penalty method nor a barrier method. Several practical applications and numerical results show that our method is strong and efficient.

[1]  David Q. Mayne,et al.  Feasible directions algorithms for optimization problems with equality and inequality constraints , 1976, Math. Program..

[2]  M. J. D. Powell,et al.  THE CONVERGENCE OF VARIABLE METRIC METHODS FOR NONLINEARLY CONSTRAINED OPTIMIZATION CALCULATIONS , 1978 .

[3]  Klaus Schittkowski,et al.  More test examples for nonlinear programming codes , 1981 .

[4]  M. J. D. Powell,et al.  Variable Metric Methods for Constrained Optimization , 1982, ISMP.

[5]  David G. Luenberger,et al.  Linear and nonlinear programming , 1984 .

[6]  José Herskovits,et al.  A two-stage feasible directions algorithm for nonlinear constrained optimization , 1981, Math. Program..

[7]  E. Panier,et al.  A superlinearly convergent feasible method for the solution of inequality constrained optimization problems , 1987 .

[8]  M. Kojima,et al.  A primal-dual interior point algorithm for linear programming , 1988 .

[9]  E. Panier,et al.  A QP-Free, globally convergent, locally superlinearly convergent algorithm for inequality constrained optimization , 1988 .

[10]  N. Megiddo Pathways to the optimal set in linear programming , 1989 .

[11]  Mauricio G. C. Resende,et al.  A Polynomial-Time Primal-Dual Affine Scaling Algorithm for Linear and Convex Quadratic Programming and Its Power Series Extension , 1990, Math. Oper. Res..

[12]  Roy E. Marsten,et al.  The interior-point method for linear programming , 1992, IEEE Software.

[13]  J. Hiriart-Urruty,et al.  Convex analysis and minimization algorithms , 1993 .

[14]  José Herskovits,et al.  An iterative algorithm for limit analysis with nonlinear yield functions , 1993 .

[15]  Jian L. Zhou,et al.  A Simple, Quadratically Convergent Interior Point Algorithm for Linear Programming and Convex Quadratic Programming , 1994 .

[16]  J. Herskovits A View on Nonlinear Optimization , 1995 .

[17]  Tao Wang,et al.  An interior point potential reduction method for constrained equations , 1996, Math. Program..

[18]  T. Tsuchiya,et al.  On the formulation and theory of the Newton interior-point method for nonlinear programming , 1996 .

[19]  José Herskovits,et al.  A quasi-Newton interior point algorithm applied to constrained optimum design in computational fluid dynamics , 1996 .

[20]  Hiroshi Yamashita A globally convergent primal-dual interior point method for constrained optimization , 1998 .

[21]  José Herskovits,et al.  Optimization theory application to slitted plate bending problems , 1998 .