Abstract
The end-to-end predict-then-optimize framework, also known as decision-focused learning, has gained popularity for its ability to integrate optimization into the training procedure of machine learning models that predict the unknown cost (objective function) coefficients of optimization problems from contextual instance information. Naturally, most of the problems of interest in this space can be cast as integer linear programs. In this work, we focus on binary linear programs (BLPs) and propose a new end-to-end training method to predict-then-optimize. Our method, Cone-aligned Vector Estimation (CaVE), aligns the predicted cost vectors with the normal cone corresponding to the true optimal solution of a training instance. When the predicted cost vector lies inside the cone, the optimal solution to the linear relaxation of the binary problem is optimal. This alignment not only produces decision-aware learning models, but also dramatically reduces training time as it circumvents the need to solve BLPs to compute a loss function with its gradients. Experiments across multiple datasets show that our method exhibits a favorable trade-off between training time and solution quality, particularly with large-scale optimization problems such as vehicle routing, a hard BLP that has yet to benefit from predict-then-optimize methods in the literature due to its difficulty.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Amos, B., Kolter, J.Z.: Optnet: differentiable optimization as a layer in neural networks. In: International Conference on Machine Learning, pp. 136–145. PMLR (2017)
Ban, G.Y., Rudin, C.: The big data newsvendor: practical insights from machine learning. Oper. Res. 67(1), 90–108 (2019)
Berthet, Q., Blondel, M., Teboul, O., Cuturi, M., Vert, J.P., Bach, F.: Learning with differentiable perturbed optimizers. arXiv preprint arXiv:2002.08676 (2020)
Dalle, G., Baty, L., Bouvier, L., Parmentier, A.: Learning with combinatorial optimization layers: a probabilistic approach. arXiv preprint arXiv:2207.13513 (2022)
Dantzig, G., Fulkerson, R., Johnson, S.: Solution of a large-scale traveling-salesman problem. J. Oper. Res. Soc. Am. 2(4), 393–410 (1954)
Elmachtoub, A., Liang, J.C.N., McNellis, R.: Decision trees for decision-making under the predict-then-optimize framework. In: International Conference on Machine Learning, vol. 119, pp. 2858–2867. PMLR (2020)
Elmachtoub, A.N., Grigas, P.: Smart predict, then optimize. Manag. Sci. (2021)
Ferber, A., Wilder, B., Dilkina, B., Tambe, M.: Mipaal: mixed integer program as a layer. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 1504–1511 (2020)
Ferber, A.M., et al.: Surco: learning linear surrogates for combinatorial nonlinear optimization problems. In: International Conference on Machine Learning, pp. 10034–10052. PMLR (2023)
Gurobi Optimization, LLC: Gurobi Optimizer Reference Manual (2021). https://www.gurobi.com
Jeong, J., Jaggi, P., Butler, A., Sanner, S.: An exact symbolic reduction of linear smart predict+ optimize to mixed integer linear programming. In: International Conference on Machine Learning, pp. 10053–10067. PMLR (2022)
Kohl, N., Desrosiers, J., Madsen, O.B., Solomon, M.M., Soumis, F.: 2-path cuts for the vehicle routing problem with time windows. Transp. Sci. 33(1), 101–116 (1999)
Liu, T.Y., et al.: Learning to rank for information retrieval. Found. Trends® Inf. Retr. 3(3), 225–331 (2009)
Mandi, J., Bucarey, V., Tchomba, M.M.K., Guns, T.: Decision-focused learning: through the lens of learning to rank. In: International Conference on Machine Learning, pp. 14935–14947. PMLR (2022)
Mandi, J., Guns, T.: Interior point solving for lp-based prediction+optimisation. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 7272–7282, Curran Associates, Inc. (2020)
Mandi, J., et al.: Decision-focused learning: foundations, state of the art, benchmark and future opportunities. arXiv preprint arXiv:2307.13565 (2023)
Mandi, J., Stuckey, P.J., Guns, T., et al.: Smart predict-and-optimize for hard combinatorial optimization problems. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 1603–1610 (2020). https://doi.org/10.1609/aaai.v34i02.5521
Miller, C.E., Tucker, A.W., Zemlin, R.A.: Integer programming formulation of traveling salesman problems. J. ACM (JACM) 7(4), 326–329 (1960)
Misra, S., Roald, L., Ng, Y.: Learning for constrained optimization: identifying optimal active constraint sets. INFORMS J. Comput. 34(1), 463–480 (2022)
Mulamba, M., Mandi, J., Diligenti, M., Lombardi, M., Bucarey, V., Guns, T.: Contrastive losses and solution caching for predict-and-optimize. arXiv preprint arXiv:2011.05354 (2020)
Niepert, M., Minervini, P., Franceschi, L.: Implicit mle: backpropagating through discrete exponential family distributions. Adv. Neural Inf. Process. Syst. 34, 14567–14579 (2021)
Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 32 (2019)
Pogančić, M.V., Paulus, A., Musil, V., Martius, G., Rolinek, M.: Differentiation of blackbox combinatorial solvers. In: International Conference on Learning Representations (2019)
Sadana, U., Chenreddy, A., Delage, E., Forel, A., Frejinger, E., Vidal, T.: A survey of contextual optimization methods for decision making under uncertainty. arXiv preprint arXiv:2306.10374 (2023)
Sahoo, S.S., Paulus, A., Vlastelica, M., Musil, V., Kuleshov, V., Martius, G.: Backpropagation through combinatorial algorithms: identity with projection works. arXiv preprint arXiv:2205.15213 (2022)
Shah, S., Perrault, A., Wilder, B., Tambe, M.: Leaving the nest: going beyond local loss functions for predict-then-optimize. arXiv preprint arXiv:2305.16830 (2023)
Shah, S., Wilder, B., Perrault, A., Tambe, M.: Learning (local) surrogate loss functions for predict-then-optimize problems. arXiv e-prints pp. arXiv–2203 (2022)
Tang, B., Khalil, E.B.: PyEPO: a pytorch-based end-to-end predict-then-optimize library for linear and integer programming. arXiv preprint arXiv:2206.14234 (2022)
Toth, P., Vigo, D.: Vehicle routing: problems, methods, and applications. SIAM (2014)
Virtanen, P., et al.: Scipy 1.0: fundamental algorithms for scientific computing in python. Nat. Methods 17(3), 261–272 (2020)
Wilder, B., Dilkina, B., Tambe, M.: Melding the data-decisions pipeline: decision-focused learning for combinatorial optimization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 1658–1665 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
A More Experiments
A More Experiments
1.1 A.1 TSP Relaxation
To enhance training efficiency, it has been proposed in [17] that a linear relaxation can be used as a substitute for the original BLP during training. Because the DFJ formulation utilizes constraint generation to handle subtour elimiation constraints, it is challenging to achieve linear relaxation due to the exponential number of constraints. In this study, we used the LP relaxation of the Miller-Tucker-Zemlin (MTZ) formulation [18] of the TSP, and trained SPO+ Rel and PFYL Rel on the same TSP instances with 20 and 50 nodes. Although relaxation methods are more efficient than CaVE in TSP20, they result in higher regret. As the size of the model increases, the relaxation approach also loses its efficiency advantage (Tables 7 and 8).
1.2 A.2 CVRP Relaxation
Similarly, the CVRP formulation with k-path cuts also struggles to obtain a linear relaxation. Thus, the MTZ formulation can be modified with constraints
for subtour elimilation and customer demand, where Q is the capacity and \(q_i\) is the demand of customer i. The SPO+ Rel and PFYL Rel are trained for the same VRP instances with 20 customers. While the linear relaxation approach demonstrates high efficiency in these scenarios, it does not guarantee a low regret compared to CaVE (Table 9).
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Tang, B., Khalil, E.B. (2024). CaVE: A Cone-Aligned Approach for Fast Predict-then-optimize with Binary Linear Programs. In: Dilkina, B. (eds) Integration of Constraint Programming, Artificial Intelligence, and Operations Research. CPAIOR 2024. Lecture Notes in Computer Science, vol 14743. Springer, Cham. https://doi.org/10.1007/978-3-031-60599-4_12
Download citation
DOI: https://doi.org/10.1007/978-3-031-60599-4_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-60601-4
Online ISBN: 978-3-031-60599-4
eBook Packages: Computer ScienceComputer Science (R0)