Convergence in Markovian models with implications for efficiency of inference

Sequential statistical models such as dynamic Bayesian networks and hidden Markov models more specifically, model stochastic processes over time. In this paper, we study for these models the effect of consecutive similar observations on the posterior probability distribution of the represented process. We show that, given such observations, the posterior distribution converges to a limit distribution. Building upon the rate of the convergence, we further show that, given some wished-for level of accuracy, part of the inference can be forestalled. To evaluate our theoretical results, we study their implications for a real-life model from the medical domain and for a benchmark model for agricultural purposes. Our results indicate that whenever consecutive similar observations arise, the computational requirements of inference in Markovian models can be drastically reduced.

[1]  Sheldon M. Ross,et al.  Stochastic Processes , 2018, Gauge Integral Structures for Stochastic Calculus and Quantum Electrodynamics.

[2]  Peter J. F. Lucas,et al.  A probabilistic and decision-theoretic approach to the management of infectious disease at the ICU , 2000, Artif. Intell. Medicine.

[3]  David J. Spiegelhalter,et al.  Probabilistic Networks and Expert Systems , 1999, Information Science and Statistics.

[4]  Niels Peek,et al.  Explicit temporal models for decision-theoretic planning of clinical management , 1999, Artif. Intell. Medicine.

[5]  Xavier Boyen,et al.  Tractable Inference for Complex Stochastic Processes , 1998, UAI.

[6]  Stuart J. Russell,et al.  Dynamic bayesian networks: representation, inference and learning , 2002 .

[7]  Lain L. MacDonald,et al.  Hidden Markov and Other Models for Discrete- valued Time Series , 1997 .

[8]  Keiji Kanazawa,et al.  A model for reasoning about persistence and causation , 1989 .

[9]  L. Baum,et al.  A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains , 1970 .

[10]  Ross D. Shachter Probabilistic Inference and Influence Diagrams , 1988, Oper. Res..

[11]  Yoshua Bengio,et al.  Input-output HMMs for sequence processing , 1996, IEEE Trans. Neural Networks.

[12]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[13]  Nir Friedman,et al.  Bayesian Network Classifiers , 1997, Machine Learning.

[14]  Pierre Baldi,et al.  Hidden Markov Models of the G-Protein-Coupled Receptor Family , 1994, J. Comput. Biol..

[15]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[16]  Peter J. F. Lucas,et al.  A dynamic Bayesian network for diagnosing ventilator-associated pneumonia in ICU patients , 2009, Expert Syst. Appl..

[17]  Uffe Kjærulff,et al.  dHugin: a computational system for dynamic time-sliced Bayesian networks , 1995 .

[18]  Anthony Unwin,et al.  Markov Chains — Theory and Applications , 1977 .

[19]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[20]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[21]  Ewart R. Carson,et al.  A Model-Based Approach to Insulin Adjustment , 1991, AIME.

[22]  Philippe Weber,et al.  Reliability modelling with dynamic bayesian networks , 2003 .