Most Relevant Explanation: computational complexity and approximation methods

Most Relevant Explanation (MRE) is the problem of finding a partial instantiation of a set of target variables that maximizes the generalized Bayes factor as the explanation for given evidence in a Bayesian network. MRE has a huge solution space and is extremely difficult to solve in large Bayesian networks. In this paper, we first prove that MRE is at least NP-hard. We then define a subproblem of MRE called MREk that finds the most relevant k-ary explanation and prove that the decision problem of MREk is $NP^{\it PP}$-complete. Since MRE needs to find the best solution by MREk over all k, and we can also show that MRE is in $NP^{\it PP}$, we conjecture that a decision problem of MRE is $NP^{\it PP}$-complete as well. Furthermore, we show that MRE remains in $NP^{\it PP}$ even if we restrict the number of target variables to be within a log factor of the number of all unobserved variables. These complexity results prompt us to develop a suite of approximation algorithms for solving MRE, One algorithm finds an MRE solution by integrating reversible-jump MCMC and simulated annealing in simulating a non-homogeneous Markov chain that eventually concentrates its mass on the mode of a distribution of the GBF scores of all solutions. The other algorithms are all instances of local search methods, including forward search, backward search, and tabu search. We tested these algorithms on a set of benchmark diagnostic Bayesian networks. Our empirical results show that these methods could find optimal MRE solutions for most of the test cases in our experiments efficiently.

[1]  Changhe Yuan Some Properties of Most Relevant Explanation , 2009, ExaCt.

[2]  Stephen A. Cook,et al.  The complexity of theorem-proving procedures , 1971, STOC.

[3]  Changhe Yuan,et al.  A General Framework for Generating Multivariate Explanations in Bayesian Networks , 2008, AAAI.

[4]  Michael L. Littman,et al.  The Computational Complexity of Probabilistic Planning , 1998, J. Artif. Intell. Res..

[5]  Changhe Yuan,et al.  Annealed MAP , 2004, UAI.

[6]  I.,et al.  Weight of Evidence : A Brief Survey , 2006 .

[7]  Tsai-Ching Lu,et al.  Focusing Strategies for Multiple Fault Diagnosis , 2006, FLAIRS.

[8]  Changhe Yuan,et al.  Most Relevant Explanation: Properties, Algorithms, and Evaluations , 2009, UAI.

[9]  Adnan Darwiche,et al.  Approximating MAP using Local Search , 2001, UAI.

[10]  Branden Fitelson,et al.  Likelihoodism, Bayesianism, and relational confirmation , 2007, Synthese.

[11]  Changhe Yuan,et al.  Most Relevant Explanation in Bayesian Networks , 2011, J. Artif. Intell. Res..

[12]  P. Green Reversible jump Markov chain Monte Carlo computation and Bayesian model determination , 1995 .

[13]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[14]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[15]  Lluís A. Belanche Muñoz,et al.  Feature selection algorithms: a survey and experimental evaluation , 2002, 2002 IEEE International Conference on Data Mining, 2002. Proceedings..

[16]  John T. Gill,et al.  Computational complexity of probabilistic Turing machines , 1974, STOC '74.

[17]  A. Darwiche,et al.  Complexity Results and Approximation Strategies for MAP Explanations , 2011, J. Artif. Intell. Res..

[18]  Branden Fitelson,et al.  STUDIES IN BAYESIAN CONFIRMATION THEORY , 2001 .

[19]  Nando de Freitas,et al.  Reversible Jump MCMC Simulated Annealing for Neural Networks , 2000, UAI.

[20]  Yan Lin,et al.  Relevance-Based Sequential Evidence Processing in Bayesian Networks , 1998, FLAIRS.