This study introduces the notion of safety for the controlled Markov chains in the continuous-time horizon. The concept is a non-trivial extension of safety control for stochastic systems modelled as discrete-time Markov decision processes, where the safety means that the probability distributions of the system states will not visit the given forbidden set at any time. In this paper study a unit-interval-valued vector that serves as an upper bound on the state probability distribution vector characterises the forbidden set. A probability distribution is then called safe if it does not exceed the upper bound. Under mild conditions the author derives two results: (i) the necessary and sufficient conditions that guarantee the all-time safety of the probability distributions if the starting distribution is safe, and (ii) the characterisation of the supreme set of safe initial probability vectors that remain safe as time passes. In particular, study the paper identifies an upper bound on time and shows that if a distribution is always safe before that time, the distribution is safe at all times. Numerical examples are provided to illustrate the two results.
[1]
E. Altman.
Constrained Markov Decision Processes
,
1999
.
[2]
Martin L. Puterman,et al.
Markov Decision Processes: Discrete Stochastic Dynamic Programming
,
1994
.
[3]
Valerie Isham,et al.
Non‐Negative Matrices and Markov Chains
,
1983
.
[4]
O. Hernández-Lerma,et al.
Continuous-time controlled Markov chains
,
2003
.
[5]
Sekhar Tangirala,et al.
Controlled Markov chains with safety upper bound
,
2003,
IEEE Trans. Autom. Control..
[6]
Ari Arapostathis,et al.
Control of Markov chains with safety bounds
,
2005,
IEEE Transactions on Automation Science and Engineering.