We consider the class of nonlinear optimal con- trol problems with all data (differential equation, state and control constraints, cost) being polynomials. We provide a simple hierarchy of LMI-relaxations whose optimal values form a nondecreasing sequence of lower bounds on the optimal value. Preliminary results show that good approximations are obtained with few moments. I. INTRODUCTION In general, solving a general nonlinear optimal control problem (OCP) is a difficult challenge, despite powerfull theoretical tools are available, e.g. the maximum principle and Hamilton-Jacobi-Bellman optimality equation. How- ever there exist many numerical methods to approximate the solution of a given optimal control problem. For in- stance, Multiple shooting techniques which solve two-point boundary value problems as decribed in e.g. (17), (7), or direct methods, as in e.g. (18), (5), (6), which for instance, use descent algorithms, among others. Contribution. In this paper, we consider the particular class of nonlinear OCPs for which all data describing the problem (dynamics, state and control constraints) are polynomials. We propose a completely different approach to provide a good approximation of (only) the optimal value of the OCP, via a sequence of increasing lower bounds. As such, it could be seen as a complement to the above shooting or direct methods which provide an upper bound, and when the sequence of lower bounds converges to the optimal value, a test of their efficiency. We first adopt an infinite-dimensional linear program- ming (LP) formulation based on the Hamilton-Jacobi- Bellman equation, as developed in e.g. Hernandez- Hernandez et al. (11). We then follow a numerical approx- imation scheme (a relaxation of the original LP) in the vein but different from the general framework developed in Hernandez and Lasserre (10) for infinite-dimensional linear programs. Our contribution is to here exploit the fact that all data are polynomials to provide a hierarchy of semidefinite programming (SDP) (or, LMI) relaxations, whose optimal values form a monotone nondecreasing sequence of lower bounds on the optimal value of the OCP. The first numerical experiments show that good approximations are obtained early in the hierarchy, i.e., with few moments, confirming
[1]
L. S. Pontryagin,et al.
Mathematical Theory of Optimal Processes
,
1962
.
[2]
M. L. Chambers.
The Mathematical Theory of Optimal Processes
,
1965
.
[3]
Philip E. Gill,et al.
Practical optimization
,
1981
.
[4]
R. W. Brockett,et al.
Asymptotic stability and feedback stabilization
,
1982
.
[5]
R. Fletcher.
Practical Methods of Optimization
,
1988
.
[6]
Oskar von Stryk,et al.
Direct and indirect methods for trajectory optimization
,
1992,
Ann. Oper. Res..
[7]
A. Bellaïche.
The tangent space in sub-riemannian geometry
,
1994
.
[8]
O. Hernández-Lerma,et al.
THE LINEAR PROGRAMMING APPROACH TO DETERMINISTIC OPTIMAL CONTROL PROBLEMS
,
1996
.
[9]
W. Grimm,et al.
Adjoint Estimation from a Direct Multiple Shooting Method
,
1997
.
[10]
W. Fleming.
Book Review: Discrete-time Markov control processes: Basic optimality criteria
,
1997
.
[11]
W. Hager,et al.
Optimal Control: Theory, Algorithms, and Applications
,
1998
.
[12]
Onésimo Hernández-Lerma,et al.
Approximation Schemes for Infinite Linear Programs
,
1998,
SIAM J. Optim..
[13]
O. Hernández-Lerma,et al.
Discrete-time Markov control processes
,
1999
.
[14]
B. Gaveau,et al.
Hamilton–Jacobi theory and the heat kernel on Heisenberg groups
,
2000
.
[15]
Jean B. Lasserre,et al.
Global Optimization with Polynomials and the Problem of Moments
,
2000,
SIAM J. Optim..
[16]
J. Stoer,et al.
Introduction to Numerical Analysis
,
2002
.
[17]
J. Lasserre.
Bounds on measures satisfying moment conditions
,
2002
.
[18]
J. Lasserre,et al.
Solving nonconvex optimization problems
,
2004,
IEEE Control Systems.
[19]
Emmanuel Trélat,et al.
Robust optimal stabilization of the Brockett integrator via a hybrid feedback
,
2005,
Math. Control. Signals Syst..