Belief Propagation and Revision in Networks with Loops

Local belief propagation rules of the sort proposed by P earl (1988) are guaranteed to converge to the optimal beliefs for singly connected networks. Recently, a n umber of researchers have empirically demonstrated good performance of these same algorithms on networks with loops, but a theoretical understanding of this performance has yet to be achieved. Here we l a y a foundation for an understanding of belief propagation in networks with loops. For networks with a single loop, we derive an analytical relationship between the steady state beliefs in the loopy network and the true posterior probability. Using this relationship we show a category of networks for which the MAP estimate obtained by belief update and by belief revision can be proven to be optimal (although the beliefs will be incorrect). We s h o w h o w nodes can use local information in the messages they receive in order to correct the steady state beliefs. Furthermore we p r o ve that for all networks with a single loop, the MAP estimate obtained by belief revision at convergence is guaranteed to give the globally optimal sequence of states. The result is independent of the length of the cycle and the size of the state space. For networks with multiple loops, we i n troduce the concept of a \balanced network" and show simulation results comparing belief revision and update in such networks. We show t h a t t h e T urbo code structure is balanced and present simulations on a toy T urbo code problem indicating the decoding obtained by belief revision at convergence is signiicantly more likely to be correct. A a b Figure 1: a. An example of the types of problems typically solved using belief propagation. Observed nodes are denoted by lled circles. A link between any t wo nodes implies a probabilistic compatability constraint. b. A simple network with a loop. Although belief propagation rules can be generalized to this network, a theoretical understanding of the algorithms behavior in such a n e t work has yet to be achieved.

[1]  A.H. Haddad,et al.  Applied optimal estimation , 1976, Proceedings of the IEEE.

[2]  Dimitri P. Bertsekas,et al.  Dynamic Programming: Deterministic and Stochastic Models , 1987 .

[3]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[4]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[5]  A. Glavieux,et al.  Near Shannon limit error-correcting coding and decoding: Turbo-codes. 1 , 1993, Proceedings of ICC '93 - IEEE International Conference on Communications.

[6]  Brendan J. Frey,et al.  Probability Propagation and Iterative Decoding , 1996 .

[7]  Dariush Divsalar,et al.  Soft-Output Decoding Algorithms in Iterative Decoding of Turbo Codes , 1996 .

[8]  Yair Weiss,et al.  Interpreting Images by Propagating Bayesian Beliefs , 1996, NIPS.

[9]  Brendan J. Frey,et al.  Bayesian networks for pattern classification, data compression, and channel coding , 1997 .

[10]  A J Viterbi,et al.  Interleaved concatenated codes: new perspectives on approaching the Shannon limit. , 1997, Proceedings of the National Academy of Sciences of the United States of America.

[11]  Michael I. Jordan,et al.  Probabilistic Independence Networks for Hidden Markov Probability Models , 1997, Neural Computation.

[12]  Jung-Fu Cheng,et al.  Turbo Decoding as an Instance of Pearl's "Belief Propagation" Algorithm , 1998, IEEE J. Sel. Areas Commun..

[13]  David J. C. MacKay,et al.  Good Error-Correcting Codes Based on Very Sparse Matrices , 1997, IEEE Trans. Inf. Theory.