An Investigation into Prediction + Optimisation for the Knapsack Problem

We study a prediction + optimisation formulation of the knapsack problem. The goal is to predict the profits of knapsack items based on historical data, and afterwards use these predictions to solve the knapsack. The key is that the item profits are not known beforehand and thus must be estimated, but the quality of the solution is evaluated with respect to the true profits. We formalise the problem, the goal of minimising expected regret and the learning problem, and investigate different machine learning approaches that are suitable for the optimisation problem. Recent methods for linear programs have incorporated the linear relaxation directly into the loss function. In contrast, we consider less intrusive techniques of changing the loss function, such as standard and multi-output regression, and learning-to-rank methods. We empirically compare the approaches on real-life energy price data and synthetic benchmarks, and investigate the merits of the different approaches.

[1]  Greet Vanden Berghe,et al.  Multi-machine energy-aware scheduling , 2017, EURO J. Comput. Optim..

[2]  Thorsten Joachims,et al.  Making large scale SVM learning practical , 1998 .

[3]  Lars Kotthoff,et al.  Algorithm Selection for Combinatorial Search Problems: A Survey , 2012, AI Mag..

[4]  James C. Spall,et al.  Introduction to stochastic search and optimization - estimation, simulation, and control , 2003, Wiley-Interscience series in discrete mathematics and optimization.

[5]  G. B. Mathews On the Partition of Numbers , 1896 .

[6]  Barry O'Sullivan,et al.  Constraint acquisition , 2017, Artif. Intell..

[7]  Adam N. Elmachtoub,et al.  Smart "Predict, then Optimize" , 2017, Manag. Sci..

[8]  Priya L. Donti,et al.  Task-based End-to-end Model Learning in Stochastic Optimization , 2017, NIPS.

[9]  Andrea Passerini,et al.  Introduction to the special issue on Combining Constraint Solving with Mining and Learning , 2017, Artif. Intell..

[10]  Andrea Passerini,et al.  Constructive Preference Elicitation by Setwise Max-Margin Learning , 2016, IJCAI.

[11]  Ralph E. Gomory,et al.  The Theory and Computation of Knapsack Functions , 1966, Oper. Res..

[12]  Andrea Passerini,et al.  Pyconstruct: Constraint Programming Meets Structured Prediction , 2018, IJCAI.

[13]  Milind Tambe,et al.  Melding the Data-Decisions Pipeline: Decision-Focused Learning for Combinatorial Optimization , 2018, AAAI.

[14]  Yuri Malitsky,et al.  DASH: Dynamic Approach for Switching Heuristics , 2013, Eur. J. Oper. Res..

[15]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[16]  Barry O'Sullivan,et al.  Analyzing the impact of electricity price forecasting on energy cost-aware scheduling , 2014, Sustain. Comput. Informatics Syst..

[17]  Concha Bielza,et al.  A survey on multi‐output regression , 2015, WIREs Data Mining Knowl. Discov..

[18]  Vladimir Vapnik,et al.  Principles of Risk Minimization for Learning Theory , 1991, NIPS.

[19]  X. Xia,et al.  Analysing the economic benefit of electricity price forecast in industrial load scheduling , 2014 .

[20]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[21]  Claude-Guy Quimper,et al.  Learning the Parameters of Global Constraints Using Branch-and-Bound , 2017, CP.

[22]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[23]  Thorsten Joachims,et al.  Training linear SVMs in linear time , 2006, KDD '06.

[24]  Eugene C. Freuder Constraint Acquisition , 2002, AISC.

[25]  Tie-Yan Liu,et al.  Learning to rank for information retrieval , 2009, SIGIR.