Parameter learning for hybrid Bayesian Networks with Gaussian mixture and Dirac mixture conditional densities

In this paper, the first algorithm for learning hybrid Bayesian Networks with Gaussian mixture and Dirac mixture conditional densities from data given their structure is presented. The mixture densities to be learned allow for nonlinear dependencies between the variables and exact closed-form inference. For learning the network's parameters, an incremental gradient ascent algorithm is derived. Analytic expressions for the partial derivatives and their combination with messages are presented. This hybrid approach subsumes the existing approach for purely discrete-valued networks and is applicable to partially observable networks, too. Its practicability is demonstrated by a reference example.

[1]  David Heckerman,et al.  A Tutorial on Learning with Bayesian Networks , 1998, Learning in Graphical Models.

[2]  P. Deb Finite Mixture Models , 2008 .

[3]  Anisse Taleb,et al.  A generic framework for blind source separation in structured nonlinear models , 2002, IEEE Trans. Signal Process..

[4]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[5]  Darryl Morrell,et al.  Implementation of Continuous Bayesian Networks Using Sums of Weighted Gaussians , 1995, UAI.

[6]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[7]  Eric Bauer,et al.  Update Rules for Parameter Estimation in Bayesian Networks , 1997, UAI.

[8]  Stuart J. Russell,et al.  Dynamic bayesian networks: representation, inference and learning , 2002 .

[9]  Robert A. Jacobs,et al.  Hierarchical Mixtures of Experts and the EM Algorithm , 1993, Neural Computation.

[10]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[11]  Stuart J. Russell,et al.  Adaptive Probabilistic Networks with Hidden Variables , 1997, Machine Learning.

[12]  Uwe D. Hanebeck,et al.  A New Approach for Hybrid Bayesian Networks Using Full Densities , 2004 .

[13]  O. C. Schrempf,et al.  Evaluation of hybrid Bayesian networks using analytical density representations , 2005 .

[14]  S. Lauritzen Propagation of Probabilities, Means, and Variances in Mixed Graphical Association Models , 1992 .

[15]  David Heckerman,et al.  A Tutorial on Learning with Bayesian Networks , 1999, Innovations in Bayesian Networks.

[16]  S. Lauritzen The EM algorithm for graphical association models with missing data , 1995 .

[17]  Steffen L. Lauritzen,et al.  Stable local computation with conditional Gaussian distributions , 2001, Stat. Comput..

[18]  Uri Lerner,et al.  Hybrid Bayesian networks for reasoning about complex systems , 2002 .