Correctness of Local Probability Propagation in Graphical Models with Loops

Graphical models, such as Bayesian networks and Markov networks, represent joint distributions over a set of variables by means of a graph. When the graph is singly connected, local propagation rules of the sort proposed by Pearl (1988) are guaranteed to converge to the correct posterior probabilities. Recently a number of researchers have empirically demonstrated good performance of these same local propagation schemes on graphs with loops, but a theoretical understanding of this performance has yet to be achieved. For graphical models with a single loop, we derive an analytical relationship between the probabilities computed using local propagation and the correct marginals. Using this relationship we show a category of graphical models with loops for which local propagation gives rise to provably optimal maximum a posteriori assignments (although the computed marginals will be incorrect). We also show how nodes can use local information in the messages they receive in order to correct their computed marginals. We discuss how these results can be extended to graphical models with multiple loops and show simulation results suggesting that some properties of propagation on single-loop graphs may hold for a larger class of graphs. Specifically we discuss the implication of our results for understanding a class of recently proposed error-correcting codes known as turbo codes.

[1]  Robert G. Gallager,et al.  Low-density parity-check codes , 1962, IRE Trans. Inf. Theory.

[2]  M. Lewin On nonnegative matrices , 1971 .

[3]  A.H. Haddad,et al.  Applied optimal estimation , 1976, Proceedings of the IEEE.

[4]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Gilbert Strang,et al.  Introduction to applied mathematics , 1988 .

[6]  Dimitri P. Bertsekas,et al.  Dynamic Programming: Deterministic and Stochastic Models , 1987 .

[7]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[8]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[9]  A. Glavieux,et al.  Near Shannon limit error-correcting coding and decoding: Turbo-codes. 1 , 1993, Proceedings of ICC '93 - IEEE International Conference on Communications.

[10]  David J. C. MacKay,et al.  Good Codes Based on Very Sparse Matrices , 1995, IMACC.

[11]  Dariush Divsalar,et al.  Soft-Output Decoding Algorithms in Iterative Decoding of Turbo Codes , 1996 .

[12]  Yair Weiss,et al.  Interpreting Images by Propagating Bayesian Beliefs , 1996, NIPS.

[13]  Niclas Wiberg,et al.  Codes and Decoding on General Graphs , 1996 .

[14]  Padhraic Smyth,et al.  Belief networks, hidden Markov models, and Markov random fields: A unifying view , 1997, Pattern Recognit. Lett..

[15]  Edward H. Adelson,et al.  Belief Propagation and Revision in Networks with Loops , 1997 .

[16]  Brendan J. Frey,et al.  Bayesian networks for pattern classification, data compression, and channel coding , 1997 .

[17]  Michael I. Jordan,et al.  Probabilistic Independence Networks for Hidden Markov Probability Models , 1997, Neural Computation.

[18]  Jung-Fu Cheng,et al.  Turbo Decoding as an Instance of Pearl's "Belief Propagation" Algorithm , 1998, IEEE J. Sel. Areas Commun..

[19]  B. Frey,et al.  Skewness and pseudocodewords in iterative decoding , 1998, Proceedings. 1998 IEEE International Symposium on Information Theory (Cat. No.98CH36252).

[20]  Brendan J. Frey,et al.  Iterative Decoding of Compound Codes by Probability Propagation in Graphical Models , 1998, IEEE J. Sel. Areas Commun..

[21]  R.J. McEliece,et al.  Iterative decoding on graphs with a single cycle , 1998, Proceedings. 1998 IEEE International Symposium on Information Theory (Cat. No.98CH36252).

[22]  Michael I. Jordan Graphical Models , 2003 .

[23]  David J. C. MacKay,et al.  Good Error-Correcting Codes Based on Very Sparse Matrices , 1997, IEEE Trans. Inf. Theory.

[24]  A. Robert Calderbank,et al.  Minimal tail-biting trellises: The Golay code and more , 1999, IEEE Trans. Inf. Theory.

[25]  Robert J. McEliece,et al.  The generalized distributive law , 2000, IEEE Trans. Inf. Theory.

[26]  D. Saad,et al.  Error-correcting code on a cactus: A solvable model , 2000, cond-mat/0005109.

[27]  William T. Freeman,et al.  Correctness of Belief Propagation in Gaussian Graphical Models of Arbitrary Topology , 1999, Neural Computation.

[28]  William T. Freeman,et al.  On the optimality of solutions of the max-product belief-propagation algorithm in arbitrary graphs , 2001, IEEE Trans. Inf. Theory.

[29]  Yair Be'ery,et al.  Convergence analysis of turbo decoding of product codes , 2001, IEEE Trans. Inf. Theory.

[30]  Amir H. Banihashemi,et al.  Decoding low-density parity-check codes with probabilistic scheduling , 2001, IEEE Communications Letters.

[31]  Brendan J. Frey,et al.  Signal-space characterization of iterative decoding , 2001, IEEE Trans. Inf. Theory.

[32]  B. Kramer,et al.  Dynamics and plasticity of peptidergic control centres in the retino‐brain‐pituitary system of Xenopus laevis , 2001, Microscopy research and technique.

[33]  Benjamin Van Roy,et al.  An analysis of belief propagation on the turbo decoding graph with Gaussian densities , 2001, IEEE Trans. Inf. Theory.

[34]  A. Willsky Multiresolution Markov models for signal and image processing , 2002, Proc. IEEE.

[35]  Yun Peng,et al.  A Neural Network Approach to Approximating Map in Belief Networks , 2002, Int. J. Neural Syst..