Correctness of Belief Propagation in Gaussian Graphical Models of Arbitrary Topology

Graphical models, such as Bayesian networks and Markov random fields, represent statistical dependencies of variables by a graph. Local belief propagation rules of the sort proposed by Pearl (1988) are guaranteed to converge to the correct posterior probabilities in singly connected graphs. Recently, good performance has been obtained by using these same rules on graphs with loops, a method we refer to as loopy belief propagation. Perhaps the most dramatic instance is the near Shannon-limit performance of Turbo codes, whose decoding algorithm is equivalent to loopy propagation. Except for the case of graphs with a single loop, there has been little theoretical understanding of loopy propagation. Here we analyze belief propagation in networks with arbitrary topologies when the nodes in the graph describe jointly gaussian random variables. We give an analytical formula relating the true posterior probabilities with those calculated using loopy propagation. We give sufficient conditions for convergence and show that when belief propagation converges, it gives the correct posterior means for all graph topologies, not just networks with a single loop. These results motivate using the powerful belief propagation algorithm in a broader class of networks and help clarify the empirical performance results.

[1]  Robert G. Gallager,et al.  Low-density parity-check codes , 1962, IRE Trans. Inf. Theory.

[2]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Gilbert Strang,et al.  Introduction to applied mathematics , 1988 .

[4]  G. B. Smith,et al.  Preface to S. Geman and D. Geman, “Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images” , 1987 .

[5]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[6]  Stig K. Andersen,et al.  Probabilistic reasoning in intelligent systems: Networks of plausible inference , 1991 .

[7]  Radford M. Neal Connectionist Learning of Belief Networks , 1992, Artif. Intell..

[8]  A. Glavieux,et al.  Near Shannon limit error-correcting coding and decoding: Turbo-codes. 1 , 1993, Proceedings of ICC '93 - IEEE International Conference on Communications.

[9]  W. C. Karl,et al.  Eecient Multiscale Regularization with Applications to the Computation of Optical Flow 1 , 1993 .

[10]  W. Clem Karl,et al.  Efficient multiscale regularization with applications to the computation of optical flow , 1994, IEEE Trans. Image Process..

[11]  Michael I. Jordan,et al.  Mean Field Theory for Sigmoid Belief Networks , 1996, J. Artif. Intell. Res..

[12]  Finn Verner Jensen,et al.  Introduction to Bayesian Networks , 2008, Innovations in Bayesian Networks.

[13]  Yair Weiss,et al.  Interpreting Images by Propagating Bayesian Beliefs , 1996, NIPS.

[14]  Niclas Wiberg,et al.  Codes and Decoding on General Graphs , 1996 .

[15]  Padhraic Smyth,et al.  Belief networks, hidden Markov models, and Markov random fields: A unifying view , 1997, Pattern Recognit. Lett..

[16]  Edward H. Adelson,et al.  Belief Propagation and Revision in Networks with Loops , 1997 .

[17]  Enrique F. Castillo,et al.  Expert Systems and Probabilistic Network Models , 1996, Monographs in Computer Science.

[18]  Brendan J. Frey,et al.  Bayesian networks for pattern classification, data compression, and channel coding , 1997 .

[19]  Jung-Fu Cheng,et al.  Turbo Decoding as an Instance of Pearl's "Belief Propagation" Algorithm , 1998, IEEE J. Sel. Areas Commun..

[20]  Michael I. Jordan,et al.  A Mean Field Learning Algorithm for Unsupervised Neural Networks , 1999, Learning in Graphical Models.

[21]  Lawrence K. Saul,et al.  Inference in Multilayer Networks via Large Deviation Bounds , 1998, NIPS.

[22]  Brendan J. Frey,et al.  Iterative Decoding of Compound Codes by Probability Propagation in Graphical Models , 1998, IEEE J. Sel. Areas Commun..

[23]  R.J. McEliece,et al.  Iterative decoding on graphs with a single cycle , 1998, Proceedings. 1998 IEEE International Symposium on Information Theory (Cat. No.98CH36252).

[24]  Michael I. Jordan Graphical Models , 2003 .

[25]  William T. Freeman,et al.  Learning to Estimate Scenes from Images , 1998, NIPS.

[26]  David Barber,et al.  Tractable Variational Structures for Approximating Graphical Models , 1998, NIPS.

[27]  Robert Cowell,et al.  Advanced Inference in Bayesian Networks , 1999, Learning in Graphical Models.

[28]  Michael I. Jordan,et al.  Loopy Belief Propagation for Approximate Inference: An Empirical Study , 1999, UAI.

[29]  Brendan J. Frey,et al.  Local Probability Propagation for Factor Analysis , 1999, NIPS.

[30]  Alan S. Willsky,et al.  The Modeling and Estimation of Statistically Self-Similar Processes in a Multiresolution Framework , 1999, IEEE Trans. Inf. Theory.

[31]  W. Freeman,et al.  Loopy Belief Propagation Gives Exact Posterior Means for Gaussian , 1999 .

[32]  Benjamin Van Roy,et al.  An Analysis of Turbo Decoding with Gaussian Densities , 1999, NIPS.

[33]  B. Frey Turbo Factor Analysis , 1999 .

[34]  Robert J. McEliece,et al.  The generalized distributive law , 2000, IEEE Trans. Inf. Theory.

[35]  Yair Weiss,et al.  Correctness of Local Probability Propagation in Graphical Models with Loops , 2000, Neural Computation.

[36]  William T. Freeman,et al.  On the fixed points of the max-product algorithm , 2000 .

[37]  William T. Freeman,et al.  On the optimality of solutions of the max-product belief-propagation algorithm in arbitrary graphs , 2001, IEEE Trans. Inf. Theory.

[38]  G. Forney,et al.  Iterative Decoding of Tail-Biting Trellises and Connections with Symbolic Dynamics , 2001 .

[39]  Benjamin Van Roy,et al.  An analysis of belief propagation on the turbo decoding graph with Gaussian densities , 2001, IEEE Trans. Inf. Theory.