Randomized dynamic programming principle and Feynman-Kac representation for optimal control of McKean-Vlasov dynamics

We analyze a stochastic optimal control problem, where the state process follows a McKean-Vlasov dynamics and the diffusion coefficient can be degenerate. We prove that its value function V admits a nonlinear Feynman-Kac representation in terms of a class of forward-backward stochastic differential equations, with an autonomous forward process. We exploit this probabilistic representation to rigorously prove the dynamic programming principle (DPP) for V. The Feynman-Kac representation we obtain has an important role beyond its intermediary role in obtaining our main result: in fact it would be useful in developing probabilistic numerical schemes for V. The DPP is important in obtaining a characterization of the value function as a solution of a non-linear partial differential equation (the so-called Hamilton-Jacobi-Belman equation), in this case on the Wasserstein space of measures. We should note that the usual way of solving these equations is through the Pontryagin maximum principle, which requires some convexity assumptions. There were attempts in using the dynamic programming approach before, but these works assumed a priori that the controls were of Markovian feedback type, which helps write the problem only in terms of the distribution of the state process (and the control problem becomes a deterministic problem). In this paper, we will consider open-loop controls and derive the dynamic programming principle in this most general case. In order to obtain the Feynman-Kac representation and the randomized dynamic programming principle, we implement the so-called randomization method, which consists in formulating a new McKean-Vlasov control problem, expressed in weak form taking the supremum over a family of equivalent probability measures. One of the main results of the paper is the proof that this latter control problem has the same value function V of the original control problem.

[1]  C. Snow,et al.  Old and new , 2017 .

[2]  Daniel Lacker,et al.  Limit Theory for Controlled McKean-Vlasov Dynamics , 2016, SIAM J. Control. Optim..

[3]  Huyên Pham,et al.  Dynamic Programming for Optimal Control of Stochastic McKean-Vlasov Dynamics , 2016, SIAM J. Control. Optim..

[4]  R. Carmona Lectures on Bsdes, Stochastic Control, and Stochastic Differential Games with Financial Applications , 2016 .

[5]  H. Pham,et al.  Randomization method and backward SDEs for optimal control of partially observed path-dependent stochastic systems , 2015, 1511.09274.

[6]  Huyen Pham,et al.  Randomized and backward SDE representation for optimal control of non-Markovian SDEs , 2015 .

[7]  A. Bensoussan,et al.  On The Interpretation Of The Master Equation , 2015, 1503.07754.

[8]  Olivier Pironneau,et al.  Dynamic Programming for Mean-Field Type Control , 2014, Journal of Optimization Theory and Applications.

[9]  Juan Li,et al.  Mean-field stochastic differential equations and associated PDEs , 2014, 1407.1215.

[10]  Andrea Cosso,et al.  Backward SDE representation for stochastic control problems with nondominated controlled intensity. , 2014, 1405.3540.

[11]  Francois Delarue,et al.  The Master Equation for Large Population Equilibriums , 2014, 1404.4694.

[12]  H. Pham,et al.  Discrete time approximation of fully nonlinear HJB equations via BSDEs with nonpositive jumps , 2013, 1311.4505.

[13]  A. Bensoussan,et al.  Mean Field Games and Mean Field Type Control Theory , 2013 .

[14]  R. Carmona,et al.  Mean Field Games and Systemic Risk , 2013, 1308.2172.

[15]  Jerzy Zabczyk,et al.  Chance and decision. Stochastic control in discrete time , 2013 .

[16]  R. Carmona,et al.  Forward-Backward Stochastic Differential Equations and Controlled McKean Vlasov Dynamics , 2013, 1303.5835.

[17]  Huyen Pham,et al.  Feynman-Kac representation for Hamilton-Jacobi-Bellman IPDE , 2012, 1212.2000.

[18]  R. Carmona,et al.  Control of McKean–Vlasov dynamics versus mean field games , 2012, 1210.5771.

[19]  Daniel Andersson,et al.  A Maximum Principle for SDEs of Mean-Field Type , 2011 .

[20]  Boualem Djehiche,et al.  A General Stochastic Maximum Principle for SDEs of Mean-field Type , 2011 .

[21]  D. Crisan,et al.  Fundamentals of Stochastic Filtering , 2008 .

[22]  Wilfrid Gangbo,et al.  Differential forms on Wasserstein space and infinite-dimensional Hamiltonian systems , 2008, 0807.1065.

[23]  Xunjing Li,et al.  Necessary Conditions for Optimal Control of Stochastic Systems with Random Jumps , 1994 .

[24]  N. Krylov Controlled Diffusion Processes , 1980 .

[25]  M. Yor,et al.  Calcul stochastique dépendant d'un paramètre , 1978 .

[26]  N. Aronszajn,et al.  EXTENSION OF UNIFORMLY CONTINUOUS TRANSFORMATIONS AND HYPERCONVEX METRIC SPACES , 1956 .

[27]  P. Cardaliaguet,et al.  Mean Field Games , 2020, Lecture Notes in Mathematics.

[28]  Marco Pavone,et al.  Stochastic Optimal Control , 2015 .

[29]  M. Yor,et al.  Continuous martingales and Brownian motion , 1990 .

[30]  M. K rn,et al.  Stochastic Optimal Control , 1988 .