Belief Optimization for Binary Networks: A Stable Alternative to Loopy Belief Propagation

We present a novel inference algorithm for arbitrary, binary, undirected graphs. Unlike loopy belief propagation, which iterates fixed point equations, we directly descend on the Bethe free energy. The algorithm consists of two phases, first we update the pairwise probabilities, given the marginal probabilities at each unit, using an analytic expression. Next, we update the marginal probabilities, by following the negative gradient of the Bethe free energy. Both steps are guaranteed to decrease the Bethe free energy, and since it is lower bounded, the algorithm is guaranteed to converge to a local minimum. We also show that the Bethe free energy is equal to the TAP free energy up to second order in the weights. In experiments we confirm that when belief propagation converges it usually finds identical solutions as our belief optimization method. The stable nature of belief optimization makes it ideally suited for learning graphical models from data.

[1]  K. Nakanishi Two- and three-spin cluster theory of spin-glasses , 1981 .

[2]  D. R. Bowman,et al.  Spin-glass theory in the Bethe approximation: Insights and problems , 1982 .

[3]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[4]  J. Yedidia,et al.  How to expand around mean-field theory using high-temperature expansions , 1991 .

[5]  Brendan J. Frey,et al.  A Revolution: Belief Propagation in Graphs with Cycles , 1997, NIPS.

[6]  Jung-Fu Cheng,et al.  Turbo Decoding as an Instance of Pearl's "Belief Propagation" Algorithm , 1998, IEEE J. Sel. Areas Commun..

[7]  Yoshiyuki Kabashima,et al.  The Belief in TAP , 1998, NIPS.

[8]  William T. Freeman,et al.  Learning to Estimate Scenes from Images , 1998, NIPS.

[9]  Michael I. Jordan,et al.  Loopy Belief Propagation for Approximate Inference: An Empirical Study , 1999, UAI.

[10]  Brendan J. Frey,et al.  Local Probability Propagation for Factor Analysis , 1999, NIPS.

[11]  Yair Weiss,et al.  Correctness of Local Probability Propagation in Graphical Models with Loops , 2000, Neural Computation.

[12]  J. Yedidia An Idiosyncratic Journey Beyond Mean Field Theory , 2000 .

[13]  W. Freeman,et al.  Generalized Belief Propagation , 2000, NIPS.

[14]  M. Opper,et al.  From Naive Mean Field Theory to the TAP Equations , 2001 .

[15]  M. Opper,et al.  An Idiosyncratic Journey Beyond Mean Field Theory , 2001 .

[16]  William T. Freeman,et al.  Correctness of Belief Propagation in Gaussian Graphical Models of Arbitrary Topology , 1999, Neural Computation.

[17]  Geoffrey E. Hinton Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.

[18]  Geoffrey E. Hinton,et al.  A New Learning Algorithm for Mean Field Boltzmann Machines , 2002, ICANN.