Learning to Reformulate for Linear Programming

It has been verified that the linear programming (LP) is able to formulate many real-life optimization problems, which can obtain the optimum by resorting to corresponding solvers such as OptVerse, Gurobi and CPLEX. In the past decades, a serial of traditional operation research algorithms have been proposed to obtain the optimum of a given LP in a fewer solving time. Recently, there is a trend of using machine learning (ML) techniques to improve the performance of above solvers. However, almost no previous work takes advantage of ML techniques to improve the performance of solver from the front end, i.e., the modeling (or formulation). In this paper, we are the first to propose a reinforcement learning-based reformulation method for LP to improve the performance of solving process. Using an open-source solver COIN-OR LP (CLP) as an environment, we implement the proposed method over two public research LP datasets and one large-scale LP dataset collected from practical production planning scenario. The evaluation results suggest that the proposed method can effectively reduce both the solving iteration number (25%↓) and the solving time (15%↓) over above datasets in average, compared to directly solving the original LP instances.

[1]  Weixiong Zhang,et al.  A Novel Local Search Algorithm for the Traveling Salesman Problem that Exploits Backbones , 2005, IJCAI.

[2]  Michael J. Todd,et al.  Mathematical programming , 2004, Handbook of Discrete and Computational Geometry, 2nd Ed..

[3]  IEEE Control Systems Letters , 2022 .

[4]  David L. Dill,et al.  Learning a SAT Solver from Single-Bit Supervision , 2018, ICLR.

[5]  Benjamin W. Wah,et al.  Wiley Encyclopedia of Computer Science and Engineering , 2009, Wiley Encyclopedia of Computer Science and Engineering.

[6]  Robert E. Bixby,et al.  Implementing the Simplex Method: The Initial Basis , 1992, INFORMS J. Comput..

[7]  George Mavrotas,et al.  Combining multiple criteria analysis, mathematical programming and Monte Carlo simulation to tackle uncertainty in Research and Development project portfolio selection: A case study from Greece , 2020, Eur. J. Oper. Res..

[8]  Navdeep Jaitly,et al.  Pointer Networks , 2015, NIPS.

[9]  Chetan Chauhan,et al.  Survey of Methods of Solving TSP along with its Implementation using Dynamic Programming Approach , 2012 .

[10]  Bruce K. Bell,et al.  Volume 5 , 1998 .

[11]  I. Maros Computational Techniques of the Simplex Method , 2002 .

[12]  Samy Bengio,et al.  Neural Combinatorial Optimization with Reinforcement Learning , 2016, ICLR.

[13]  Abolfazl Gharaei,et al.  Joint Economic Lot-sizing in Multi-product Multi-level Integrated Supply Chains: Generalized Benders Decomposition , 2020, International Journal of Systems Science: Operations & Logistics.

[14]  Marco Furini,et al.  International Journal of Computer and Applications , 2010 .

[15]  Yoshua Bengio,et al.  Hybrid Models for Learning to Branch , 2020, NeurIPS.

[16]  Maria Prandini,et al.  Hyper-Graph Partitioning for a Multi-Agent Reformulation of Large-Scale MILPs , 2021, IEEE Control Systems Letters.

[17]  Erling D. Andersen,et al.  On implementing a primal-dual interior-point method for conic quadratic optimization , 2003, Math. Program..

[18]  Maria-Florina Balcan,et al.  Learning to Branch , 2018, ICML.

[19]  Andrea Lodi,et al.  Exact Combinatorial Optimization with Graph Convolutional Neural Networks , 2019, NeurIPS.