Optimal control of stochastic systems with costly observations - the general Markovian model and the LQG problem

In this paper, we examine a discrete-time stochastic control problem in which there are a number of observation options available to the controller, with varying associated costs. The observation costs are added to the running cost of the optimization criterion and the resulting optimal control problem is investigated. This problem is motivated by the wide deployment of networked control systems and data fusion. Since only part of the observation information is available at each time step, the controller has to balance the system performance with the penalty of the requested information (query). We first formulate the problem for a general partially observed Markov decision process (POMDP) model and then specialize to the stochastic LQG problem, where we show that the separation principle still holds. Moreover we show that the effect of the observation cost is manifested on the estimation strategy as follows: instead of a Kalman filter with gain determined by the algebraic Riccati equation, the optimal estimator includes, in addition, a query strategy which is characterized by a dynamic programming equation. The structure of the optimal query for a one-dimensional system is studied analytically and simulated with numerical examples.

[1]  Robin J. Evans,et al.  The problem of optimal robust sensor scheduling , 2000, Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187).

[2]  Sekhar Tatikonda,et al.  Control under communication constraints , 2004, IEEE Transactions on Automatic Control.

[3]  Dimitri P. Bertsekas,et al.  Dynamic Programming: Deterministic and Stochastic Models , 1987 .

[4]  João Pedro Hespanha,et al.  Stabilization of nonlinear systems with limited information feedback , 2005, IEEE Transactions on Automatic Control.

[5]  D. Williams CONTROLLED MARKOV PROCESSES (Grundlehren der mathematischen Wissenschaften, 235) , 1982 .

[6]  Nicola Elia,et al.  Stabilization of linear systems with limited information , 2001, IEEE Trans. Autom. Control..

[7]  Sekhar Tatikonda,et al.  Control over noisy channels , 2004, IEEE Transactions on Automatic Control.

[8]  S. Eddy Hidden Markov models. , 1996, Current opinion in structural biology.

[9]  M. K. Ghosh,et al.  Discrete-time controlled Markov processes with average cost criterion: a survey , 1993 .

[10]  Robin J. Evans,et al.  Hidden Markov model multiarm bandits: a methodology for beam scheduling in multitarget tracking , 2001, IEEE Trans. Signal Process..

[11]  Onésimo Hernández-Lerma,et al.  Controlled Markov Processes , 1965 .

[12]  R. Evans,et al.  Stabilization with data-rate-limited feedback: tightest attainable bounds , 2000 .

[13]  John B. Moore,et al.  Hidden Markov Models: Estimation and Control , 1994 .

[14]  R. Brockett,et al.  Systems with finite communication bandwidth constraints. I. State estimation problems , 1997, IEEE Trans. Autom. Control..

[15]  Robin J. Evans,et al.  The problem of optimal robust sensor scheduling , 2001, Syst. Control. Lett..

[16]  Wing Shing Wong,et al.  Systems with finite communication bandwidth constraints. II. Stabilization with limited information feedback , 1999, IEEE Trans. Autom. Control..

[17]  Daniel Liberzon,et al.  On stabilization of linear systems with limited information , 2003, IEEE Trans. Autom. Control..

[18]  Robin J. Evans,et al.  Topological feedback entropy and Nonlinear stabilization , 2004, IEEE Transactions on Automatic Control.