Inference and Learning in Hybrid Bayesian Networks

We survey the literature on methods for inference and learning in Bayesian Networks composed of discrete and continuous nodes, in which the continuous nodes have a multivariate Gaussian distribution, whose mean and variance depends on the values of the discrete nodes. We also briefly consider hybrid Dynamic Bayesian Networks, an extension of switching Kalman filters. This report is meant to summarize what is known at a sufficient level of detail to enable someone to implement the algorithms, but without dwelling on formalities.

[1]  Radford M. Neal A new view of the EM algorithm that justifies incremental and other variants , 1993 .

[2]  Yaakov Bar-Shalom,et al.  Estimation and Tracking: Principles, Techniques, and Software , 1993 .

[3]  Chang‐Jin Kim,et al.  Dynamic linear models with Markov-switching , 1994 .

[4]  Xavier Boyen,et al.  Tractable Inference for Complex Stochastic Processes , 1998, UAI.

[5]  Xavier Boyen,et al.  Approximate Learning of Dynamic Models , 1998, NIPS.

[6]  Kristian G. Olesen,et al.  Causal Probabilistic Networks with Both Discrete and Continuous Variables , 1993, IEEE Trans. Pattern Anal. Mach. Intell..

[7]  Michael I. Jordan,et al.  Probabilistic Independence Networks for Hidden Markov Probability Models , 1997, Neural Computation.

[8]  Hong Xu,et al.  Computing Marginals for Arbitrary Subsets from Marginal Representation in Markov Trees , 1995, Artif. Intell..

[9]  Kevin P. Murphy,et al.  A Variational Approximation for Bayesian Networks with Discrete and Continuous Latent Variables , 1999, UAI.

[10]  G. McLachlan,et al.  The EM algorithm and extensions , 1996 .

[11]  Darryl Morrell,et al.  Implementation of Continuous Bayesian Networks Using Sums of Weighted Gaussians , 1995, UAI.

[12]  Frank Jensen,et al.  From Influence Diagrams to junction Trees , 1994, UAI.

[13]  Michael I. Jordan Graphical Models , 2003 .

[14]  Ross D. Shachter,et al.  Fusion and Propagation with Multiple Observations in Belief Networks , 1991, Artif. Intell..

[15]  M. V. Rossum,et al.  In Neural Computation , 2022 .

[16]  Wray L. Buntine Operations for Learning with Graphical Models , 1994, J. Artif. Intell. Res..

[17]  Zoubin Ghahramani,et al.  A Unifying Review of Linear Gaussian Models , 1999, Neural Computation.

[18]  Y. Bar-Shalom Tracking and data association , 1988 .

[19]  M. Degroot Optimal Statistical Decisions , 1970 .

[20]  Kuo-Chu Chang,et al.  Symbolic Probabilistic Inference with Continuous Variables , 1994, UAI.

[21]  Michael I. Jordan,et al.  On Convergence Properties of the EM Algorithm for Gaussian Mixtures , 1996, Neural Computation.

[22]  Kuo-Chu Chang,et al.  Symbolic probabilistic inference with both discrete and continuous variables , 1995, IEEE Trans. Syst. Man Cybern..

[23]  N. Wermuth,et al.  Graphical Models for Associations between Variables, some of which are Qualitative and some Quantitative , 1989 .

[24]  Geoffrey E. Hinton,et al.  Switching State-Space Models , 1996 .

[25]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[26]  R. Redner,et al.  Mixture densities, maximum likelihood, and the EM algorithm , 1984 .

[27]  S. Lauritzen Propagation of Probabilities, Means, and Variances in Mixed Graphical Association Models , 1992 .

[28]  Michael P. Wellman,et al.  Planning and Control , 1991 .

[29]  David Heckerman,et al.  Learning Gaussian Networks , 1994, UAI.

[30]  David J. Spiegelhalter,et al.  Local computations with probabilities on graphical structures and their application to expert systems , 1990 .

[31]  Zoubin Ghahramani,et al.  Learning Dynamic Bayesian Networks , 1997, Summer School on Neural Networks.

[32]  Alice M. Agogino,et al.  Inference Using Message Propagation and Topology Transformation in Vector Gaussian Continuous Networks , 1996, UAI.