Stochastic output-feedback model predictive control

A new formulation of Stochastic Model Predictive Output Feedback Control is presented and analyzed as a transposition of Stochastic Optimal Output Feedback Control into a receding horizon setting. This requires lifting the design into a framework involving propagation of the conditional state density, the information state, and solution of the Stochastic Dynamic Programming Equation for an optimal feedback policy, both stages of which are computationally challenging in the general, nonlinear setup. The upside is that the clearance of three bottleneck aspects of Model Predictive Control is connate to the optimality: output feedback is incorporated naturally; dual regulation and probing of the control signal is inherent; closed-loop performance relative to infinite-horizon optimal control is guaranteed. While the methods are numerically formidable, our aim is to develop an approach to Stochastic Model Predictive Control with guarantees and, from there, to seek a less onerous approximation. To this end, we discuss in particular the class of Partially Observable Markov Decision Processes, to which our results extend seamlessly, and demonstrate applicability with an example in healthcare decision making, where duality and associated optimality in the control signal are required for satisfactory closed-loop behavior.

[1]  J. Doob Stochastic processes , 1953 .

[2]  L. Grüne,et al.  Nonlinear Model Predictive Control : Theory and Algorithms. 2nd Edition , 2011 .

[3]  H. Genceli,et al.  New approach to constrained predictive control with simultaneous model identification , 1996 .

[4]  A. Mesbah,et al.  Stochastic Model Predictive Control: An Overview and Perspectives for Future Research , 2016, IEEE Control Systems.

[5]  Robert R. Bitmead,et al.  Tractable dual optimal stochastic model predictive control: An example in healthcare , 2017, 2017 IEEE Conference on Control Technology and Applications (CCTA).

[6]  João Pedro Hespanha,et al.  Nonlinear output-feedback model predictive control with moving horizon estimation , 2014, 53rd IEEE Conference on Decision and Control.

[7]  Lars Grüne,et al.  On the Infinite Horizon Performance of Receding Horizon Controllers , 2008, IEEE Transactions on Automatic Control.

[8]  Dan Simon,et al.  Optimal State Estimation: Kalman, H∞, and Nonlinear Approaches , 2006 .

[9]  Vernor Vinge,et al.  A Fire Upon the Deep , 1992 .

[10]  David Q. Mayne,et al.  Model predictive control: Recent developments and future promise , 2014, Autom..

[11]  Robin J. Evans,et al.  Hidden Markov model multiarm bandits: a methodology for beam scheduling in multitarget tracking , 2001, IEEE Trans. Signal Process..

[12]  Robert R. Bitmead,et al.  Particle Model Predictive Control: Tractable Stochastic Nonlinear Output-Feedback MPC , 2016, 1612.00505.

[13]  Robert R. Bitmead,et al.  Performance of Model Predictive Control of POMDPs , 2017, 2018 European Control Conference (ECC).

[14]  Michael Nikolaou,et al.  Chance‐constrained model predictive control , 1999 .

[15]  Robert R. Bitmead,et al.  Markov modeling in hepatitis B screening and linkage to care , 2017, Theoretical Biology and Medical Modelling.

[16]  M. Hovd,et al.  Robust output feedback model predictive control for linear systems via moving horizon estimation , 2008, 2008 American Control Conference.

[17]  Robert R. Bitmead,et al.  Sumptus cohiberi: The cost of constraints in MPC with state estimates , 2016, 2016 American Control Conference (ACC).

[18]  David Q. Mayne,et al.  Robust output feedback model predictive control of constrained linear systems: Time varying case , 2009, Autom..

[19]  O. Hernández-Lerma,et al.  Error bounds for rolling horizon policies in discrete-time Markov control processes , 1990 .

[20]  W. Rudin Real and complex analysis , 1968 .

[21]  Robin J. Evans,et al.  Simulation-Based Optimal Sensor Scheduling with Application to Observer Trajectory Planning , 2005, Proceedings of the 44th IEEE Conference on Decision and Control.

[22]  Basil Kouvaritakis,et al.  Model Predictive Control , 2016 .

[23]  S. Ponomarev Submersions and preimages of sets of measure zero , 1987 .

[24]  Pravin Varaiya,et al.  Stochastic Systems: Estimation, Identification, and Adaptive Control , 1986 .

[25]  Robert R. Bitmead,et al.  Persistently exciting model predictive control , 2014 .

[26]  John Lygeros,et al.  On Stability and Performance of Stochastic Predictive Control Techniques , 2013, IEEE Transactions on Automatic Control.

[27]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[28]  D. Naidu,et al.  Optimal Control Systems , 2018 .

[29]  Suman Chakravorty,et al.  Information Space Receding Horizon Control , 2013, IEEE Transactions on Cybernetics.

[30]  Robert R. Bitmead,et al.  MPC under the hood/sous le capot/unter der Haube , 2012 .

[31]  Jun Yan,et al.  Incorporating state estimation into model predictive control and its application to network traffic control , 2005, Autom..

[32]  J. Doob Classical potential theory and its probabilistic counterpart , 1984 .

[33]  Mark Cannon,et al.  Stochastic model predictive control , 2015 .

[34]  Graham C. Goodwin,et al.  Robust model predictive control: reflections and opportunities , 2014 .

[35]  Dimitri P. Bertsekas,et al.  Stochastic optimal control : the discrete time case , 2007 .

[36]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .