Solving Constrained Combinatorial Optimisation Problems via MAP Inference without High-Order Penalties

Solving constrained combinatorial optimization problems via MAP inference is often achieved by introducing extra potential functions for each constraint. This can result in very high order potentials, e.g. a 2-order objective with pairwise potentials and a quadratic constraint over all N variables would correspond to an unconstrained objective with an order-N potential. This limits the practicality of such an approach, since inference with high order potentials is tractable only for a few special classes of functions. We propose an approach which is able to solve constrained combinatorial problems using belief propagation without increasing the order. For example, in our scheme the 2-order problem above remains order 2 instead of order N . Experiments on applications ranging from foreground detection, image reconstruction, quadratic knapsack, and the M-best solutions problem demonstrate the effectiveness and efficiency of our method. Moreover, we show several situations in which our approach outperforms commercial solvers like CPLEX and others designed for specific constrained MAP inference problems. Introduction Maximum a posteriori (MAP) inference for graphical models can be used to solve unconstrained combinatorial optimization problems, or constrained problems by introducing extra potential functions for each constraint (Ravanbakhsh, Rabbany, and Greiner 2014; Ravanbakhsh and Greiner 2014; Frey and Dueck 2007; Bayati, Shah, and Sharma 2005; Werner 2008). The main limitation of this approach is that it often results in very high order potentials. Problems with pairwise potentials, for example, are very common, and adding a quadratic constraint (order 2) over N variables results in an objective function of order N . Optimizing over such high-order potentials is tractable only for a few special classes of functions (Tarlow, Givoni, and Zemel 2010; Potetz and Lee 2008; Komodakis and Paragios 2009; Mézard, Parisi, and Zecchina 2002; Aguiar et al. 2011), such as linear functions. Recently, Lim, Jung, and Kohli (2014) proposed cuttingplane based methods to handle constrained problems without Copyright c © 2017, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. introducing very high order potentials. However, their approaches require exact solutions of a series of unconstrained MAP inference problems, which is, in general, intractable. Thus their approaches are again only applicable to a particular class of potentials and constraints. To tackle general constrained combinatorial problems, our overall idea is to formulate the unconstrained combinatorial problem as a linear program (LP) with local marginal polytope constraints only (which corresponds to a classical MAP inference problem), and then add the “real” constraints from the original combinatorial problem to the existing LP to form a new LP. Duality of the new LP absorbs the “real” constraints naturally, and yields a convenient message passing procedure. The proposed algorithm is guaranteed to find feasible solutions for a quite general set of constraints. We apply our method to problems including foreground detection, image reconstruction, quadratic knapsack, and the M-best solutions problem, and show several situations in which it outperforms the commercial optimization solver CPLEX. We also test our method against more restrictive approaches including Aguiar et al. (2011) and Lim, Jung, and Kohli (2014) on the subsets of our applications to which they are applicable. Our method outperforms these methods in most cases even in settings that favor them. Preliminaries Here we consider factor graphical models with discrete variables. Denote the graph G = (V,C), where V is the set of nodes, and C is a collection of subsets of V. Each c ∈ C is called a cluster. If we associate one random variable xi with each node, and let x = [xi]i∈V, then it is often assumed that the joint distribution of x belongs to the exponential family p(x) = 1 Z exp [∑ c∈C θc(xc) ] , where xc denotes the vector [xi]i∈c. The real-valued function θc(xc) is known as a potential function. Without loss of generality we make the following assumption to simplify the derivation: Assumption 1. For convenience we assume that: (1) C includes every node, i.e., ∀i ∈ V, {i} ∈ C; (2) C is closed under intersection, i.e., ∀c1, c2 ∈ C, it is true that c1 ∩ c2 ∈ C. MAP and its LP relaxations The goal of MAP inference is to find the most likely assignment of values to the random Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17)

[1]  Fred W. Glover,et al.  A computational study on the quadratic knapsack problem with multiple constraints , 2012, Comput. Oper. Res..

[2]  Solomon Eyal Shimony,et al.  Finding MAPs for Belief Networks is NP-Hard , 1994, Artif. Intell..

[3]  Gregory Shakhnarovich,et al.  Diverse M-Best Solutions in Markov Random Fields , 2012, ECCV.

[4]  D. Sontag 1 Introduction to Dual Decomposition for Inference , 2010 .

[5]  Anna C. Gilbert,et al.  Sparse Gradient Image Reconstruction Done Faster , 2007, 2007 IEEE International Conference on Image Processing.

[6]  Marc Pollefeys,et al.  Globally Convergent Dual MAP LP Relaxation Solvers using Fenchel-Young Margins , 2012, NIPS.

[7]  Thomas Hofmann,et al.  Using Combinatorial Optimization within Max-Product Belief Propagation , 2007 .

[8]  Tommi S. Jaakkola,et al.  Convergence Rate Analysis of MAP Coordinate Minimization Algorithms , 2012, NIPS.

[9]  Rainer Stiefelhagen,et al.  Improving foreground segmentations with probabilistic superpixel Markov random fields , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[10]  Kyomin Jung,et al.  Efficient Energy Minimization for Enforcing Label Statistics , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Silvio Savarese,et al.  Efficient and Exact MAP-MRF Inference using Branch and Bound , 2012, AISTATS.

[12]  Yair Weiss,et al.  Linear Programming Relaxations and Belief Propagation - An Empirical Study , 2006, J. Mach. Learn. Res..

[13]  Devavrat Shah,et al.  Maximum weight matching via max-product belief propagation , 2005, Proceedings. International Symposium on Information Theory, 2005. ISIT 2005..

[14]  M. Mézard,et al.  Analytic and Algorithmic Solution of Random Satisfiability Problems , 2002, Science.

[15]  Tai Sing Lee,et al.  Efficient belief propagation for higher-order cliques using linear constraint nodes , 2008, Comput. Vis. Image Underst..

[16]  Amir Globerson,et al.  An LP View of the M-best MAP problem , 2009, NIPS.

[17]  Stan Z. Li Markov Random Field Modeling in Image Analysis , 2009, Advances in Pattern Recognition.

[18]  Vladimir Kolmogorov,et al.  Convergent Tree-Reweighted Message Passing for Energy Minimization , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Ofer Meshi,et al.  An Alternating Direction Method for Dual MAP LP Relaxation , 2011, ECML/PKDD.

[20]  Nikos Komodakis,et al.  Beyond pairwise energies: Efficient optimization for higher-order MRFs , 2009, CVPR.

[21]  Vladimir Kolmogorov,et al.  An Analysis of Convex Relaxations for MAP Estimation , 2007, NIPS.

[22]  Fatih Murat Porikli,et al.  Changedetection.net: A new change detection benchmark dataset , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[23]  Russell Greiner,et al.  Augmentative Message Passing for Traveling Salesman Problem and Graph Partitioning , 2014, NIPS.

[24]  Tomás Werner,et al.  High-arity interactions, polyhedral relaxations, and cutting plane algorithm for soft constraint optimisation (MAP-MRF) , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[25]  Dhruv Batra,et al.  An Efficient Message-Passing Algorithm for the M-Best MAP Problem , 2012, UAI.

[26]  Delbert Dueck,et al.  Clustering by Passing Messages Between Data Points , 2007, Science.

[27]  H. Kellerer,et al.  Introduction to NP-Completeness of Knapsack Problems , 2004 .

[28]  Richard S. Zemel,et al.  HOP-MAP: Efficient Message Passing with High Order Potentials , 2010, AISTATS.

[29]  Tommi S. Jaakkola,et al.  Fixing Max-Product: Convergent Message Passing Algorithms for MAP LP-Relaxations , 2007, NIPS.

[30]  Russell Greiner,et al.  Perturbed message passing for constraint satisfaction problems , 2014, J. Mach. Learn. Res..

[31]  Tommi S. Jaakkola,et al.  Tightening LP Relaxations for MAP using Message Passing , 2008, UAI.

[32]  Eric P. Xing,et al.  An Augmented Lagrangian Approach to Constrained MAP Inference , 2011, ICML.