Melding the Data-Decisions Pipeline: Decision-Focused Learning for Combinatorial Optimization

Creating impact in real-world settings requires artificial intelligence techniques to span the full pipeline from data, to predictive models, to decisions. These components are typically approached separately: a machine learning model is first trained via a measure of predictive accuracy, and then its predictions are used as input into an optimization algorithm which produces a decision. However, the loss function used to train the model may easily be misaligned with the end goal, which is to make the best decisions possible. Hand-tuning the loss function to align with optimization is a difficult and error-prone process (which is often skipped entirely).We focus on combinatorial optimization problems and introduce a general framework for decision-focused learning, where the machine learning model is directly trained in conjunction with the optimization algorithm to produce highquality decisions. Technically, our contribution is a means of integrating common classes of discrete optimization problems into deep learning or other predictive models, which are typically trained via gradient descent. The main idea is to use a continuous relaxation of the discrete problem to propagate gradients through the optimization procedure. We instantiate this framework for two broad classes of combinatorial problems: linear programs and submodular maximization. Experimental results across a variety of domains show that decisionfocused learning often leads to improved optimization performance compared to traditional methods. We find that standard measures of accuracy are not a reliable proxy for a predictive model’s utility in optimization, and our method’s ability to specify the true goal as the model’s training objective yields substantial dividends across a range of decision problems.

[1]  Rishabh K. Iyer,et al.  Monotone Closure of Relaxed Constraints in Submodular Optimization: Connections Between Minimization and Maximization , 2014, UAI.

[2]  Stefanie Jegelka,et al.  Submodular meets Structured: Finding Diverse Subsets in Exponentially-Large Structured Item Sets , 2014, NIPS.

[3]  Dimitris Bertsimas,et al.  Optimal classification trees , 2017, Machine Learning.

[4]  Bo An,et al.  Deploying PAWS: Field Optimization of the Protection Assistant for Wildlife Security , 2016, AAAI.

[5]  Priya L. Donti,et al.  Task-based End-to-end Model Learning in Stochastic Optimization , 2017, NIPS.

[6]  A. Greenberg,et al.  Cope: Traffic Engineering in Dynamic Networks , 2022 .

[7]  Milind Tambe,et al.  Beware the Soothsayer: From Attack Prediction Accuracy to Predictive Reliability in Security Games , 2015, GameSec.

[8]  Gautam Biswas,et al.  Prioritized Allocation of Emergency Responders based on a Continuous-Time Incident Prediction Model , 2017, AAMAS.

[9]  George L. Nemhauser,et al.  Learning to Run Heuristics in Tree Search , 2017, IJCAI.

[10]  Andreas Krause,et al.  Differentiable Learning of Submodular Models , 2017, NIPS 2017.

[11]  Carla P. Gomes,et al.  Avicaching: A Two Stage Game for Bias Reduction in Citizen Science , 2016, AAMAS.

[12]  Éva Tardos,et al.  Maximizing the Spread of Influence through a Social Network , 2015, Theory Comput..

[13]  Hiroya Takamura,et al.  Text Summarization Model based on Maximum Coverage Problem and its Variant , 2008 .

[14]  Umesh Bellur,et al.  Improved Matchmaking Algorithm for Semantic Web Services Based on Bipartite Graph Matching , 2007, IEEE International Conference on Web Services (ICWS 2007).

[15]  Takuro Fukunaga,et al.  Threshold Influence Model for Allocating Advertising Budgets , 2015, ICML.

[16]  Andreas Krause,et al.  Guaranteed Non-convex Optimization: Submodular Maximization over Continuous Domains , 2016, AISTATS.

[17]  E. Horvitz From Data to Predictions and Decisions: Enabling Evidence-Based Healthcare , 2010 .

[18]  John Langford,et al.  The offset tree for learning with partial labels , 2008, KDD.

[19]  M. R. Rao,et al.  Combinatorial Optimization , 1992, NATO ASI Series.

[20]  Craig Boutilier,et al.  Optimal Bayesian Recommendation Sets and Myopically Optimal Choice Query Sets , 2010, NIPS.

[21]  Noga Alon,et al.  Optimizing budget allocation among channels and influencers , 2012, WWW.

[22]  Zheng Wen,et al.  Optimal Greedy Diversity for Recommendation , 2015, IJCAI.

[23]  Andreas Krause,et al.  Differentiable Submodular Maximization , 2018, IJCAI.

[24]  Michael I. Jordan,et al.  How to Escape Saddle Points Efficiently , 2017, ICML.

[25]  Claire Cardie,et al.  SparseMAP: Differentiable Sparse Structured Inference , 2018, ICML.

[26]  Lifu Tu,et al.  Learning Approximate Inference Networks for Structured Prediction , 2018, ICLR.

[27]  Adam N. Elmachtoub,et al.  Smart "Predict, then Optimize" , 2017, Manag. Sci..

[28]  Amin Karbasi,et al.  Gradient Methods for Submodular Maximization , 2017, NIPS.

[29]  Justin Domke,et al.  Generic Methods for Optimization-Based Modeling , 2012, AISTATS.

[30]  Yair Zick,et al.  Diversity Constraints in Public Housing Allocation , 2017, AAMAS.

[31]  Le Song,et al.  2 Common Formulation for Greedy Algorithms on Graphs , 2018 .

[32]  Anoop Cherian,et al.  On Differentiating Parameterized Argmin and Argmax Problems with Application to Bi-level Optimization , 2016, ArXiv.

[33]  Jan Vondrák,et al.  Maximizing a Monotone Submodular Function Subject to a Matroid Constraint , 2011, SIAM J. Comput..

[34]  Navdeep Jaitly,et al.  Pointer Networks , 2015, NIPS.

[35]  Lise Getoor,et al.  Collective Classification in Network Data , 2008, AI Mag..

[36]  Ken-ichi Kawarabayashi,et al.  Optimal Budget Allocation: Theoretical Guarantee and Efficient Algorithm , 2014, ICML.

[37]  Andrew McCallum,et al.  End-to-End Learning for Structured Prediction Energy Networks , 2017, ICML.

[38]  Vipin Kumar,et al.  A Fast and High Quality Multilevel Scheme for Partitioning Irregular Graphs , 1998, SIAM J. Sci. Comput..

[39]  Sreenivas Gollapudi,et al.  Diversifying search results , 2009, WSDM '09.

[40]  Eric Horvitz,et al.  From Data to Knowledge to Action: A Global Enabler for the 21st Century , 2020, ArXiv.

[41]  J. Zico Kolter,et al.  OptNet: Differentiable Optimization as a Layer in Neural Networks , 2017, ICML.