On Lifting the Gibbs Sampling Algorithm

First-order probabilistic models combine the power of first-order logic, the de facto tool for handling relational structure, with probabilistic graphical models, the de facto tool for handling uncertainty. Lifted probabilistic inference algorithms for them have been the subject of much recent research. The main idea in these algorithms is to improve the accuracy and scalability of existing graphical models' inference algorithms by exploiting symmetry in the first-order representation. In this paper, we consider blocked Gibbs sampling, an advanced MCMC scheme, and lift it to the first-order level. We propose to achieve this by partitioning the first-order atoms in the model into a set of disjoint clusters such that exact lifted inference is polynomial in each cluster given an assignment to all other atoms not in the cluster. We propose an approach for constructing the clusters and show how it can be used to trade accuracy with computational complexity in a principled manner. Our experimental evaluation shows that lifted Gibbs sampling is superior to the propositional algorithm in terms of accuracy, scalability and convergence.

[1]  Matthew Richardson,et al.  Markov logic networks , 2006, Machine Learning.

[2]  D. Rubin,et al.  Inference from Iterative Simulation Using Multiple Sequences , 1992 .

[3]  Mathias Niepert,et al.  Markov Chains on Orbits of Permutation Groups , 2012, UAI.

[4]  Bart Selman,et al.  Accelerated Adaptive Markov Chain for Partition Function Computation , 2011, NIPS.

[5]  Uffe Kjærulff,et al.  Blocking Gibbs sampling in very large probabilistic expert systems , 1995, Int. J. Hum. Comput. Stud..

[6]  Pedro M. Domingos,et al.  Lifted First-Order Belief Propagation , 2008, AAAI.

[7]  Stochastic Relaxation , 2014, Computer Vision, A Reference Guide.

[8]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[9]  Pedro M. Domingos,et al.  A General Method for Reducing the Complexity of Relational Inference and its Application to MCMC , 2008, AAAI.

[10]  Ben Taskar,et al.  Introduction to Statistical Relational Learning (Adaptive Computation and Machine Learning) , 2007 .

[11]  Ben Taskar,et al.  Introduction to statistical relational learning , 2007 .

[12]  Dan Suciu,et al.  Lifted Inference Seen from the Other Side : The Tractable Features , 2010, NIPS.

[13]  Pedro M. Domingos,et al.  Markov Logic: An Interface Layer for Artificial Intelligence , 2009, Markov Logic: An Interface Layer for Artificial Intelligence.

[14]  Radford M. Neal Slice Sampling , 2003, The Annals of Statistics.

[15]  John W. Lloyd,et al.  Probabilistic modelling, inference and learning using logical theories , 2008, Annals of Mathematics and Artificial Intelligence.

[16]  Matthew Richardson,et al.  The Alchemy System for Statistical Relational AI: User Manual , 2007 .

[17]  Jun S. Liu,et al.  Monte Carlo strategies in scientific computing , 2001 .

[18]  Stuart J. Russell,et al.  General-Purpose MCMC Inference over Relational Structures , 2006, UAI.

[19]  Jun S. Liu,et al.  Covariance structure of the Gibbs sampler with applications to the comparisons of estimators and augmentation schemes , 1994 .

[20]  Stig K. Andersen,et al.  Probabilistic reasoning in intelligent systems: Networks of plausible inference , 1991 .

[21]  Dan Roth,et al.  Lifted First-Order Probabilistic Inference , 2005, IJCAI.

[22]  Pedro M. Domingos,et al.  Sound and Efficient Inference with Probabilistic and Deterministic Dependencies , 2006, AAAI.

[23]  Donald Geman,et al.  Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images , 1984 .

[24]  Pedro M. Domingos,et al.  Recursive Random Fields , 2007, IJCAI.

[25]  Vibhav Gogate,et al.  Advances in Lifted Importance Sampling , 2012, AAAI.

[26]  Henry A. Kautz,et al.  Solving Bayesian Networks by Weighted Model Counting , 2005 .

[27]  Isi Mitrani,et al.  Probabilistic Modelling , 1998 .

[28]  Leslie Pack Kaelbling,et al.  Lifted Probabilistic Inference with Counting Formulas , 2008, AAAI.

[29]  Adnan Darwiche,et al.  On probabilistic inference by weighted model counting , 2008, Artif. Intell..

[30]  Michael I. Jordan,et al.  Loopy Belief Propagation for Approximate Inference: An Empirical Study , 1999, UAI.

[31]  David Poole,et al.  First-order probabilistic inference , 2003, IJCAI.

[32]  Dan Klein,et al.  Type-Based MCMC , 2010, HLT-NAACL.