CaVE: A Cone-Aligned Approach for Fast Predict-then-optimize with Binary Linear Programs | SpringerLink
Skip to main content

CaVE: A Cone-Aligned Approach for Fast Predict-then-optimize with Binary Linear Programs

  • Conference paper
  • First Online:
Integration of Constraint Programming, Artificial Intelligence, and Operations Research (CPAIOR 2024)

Abstract

The end-to-end predict-then-optimize framework, also known as decision-focused learning, has gained popularity for its ability to integrate optimization into the training procedure of machine learning models that predict the unknown cost (objective function) coefficients of optimization problems from contextual instance information. Naturally, most of the problems of interest in this space can be cast as integer linear programs. In this work, we focus on binary linear programs (BLPs) and propose a new end-to-end training method to predict-then-optimize. Our method, Cone-aligned Vector Estimation (CaVE), aligns the predicted cost vectors with the normal cone corresponding to the true optimal solution of a training instance. When the predicted cost vector lies inside the cone, the optimal solution to the linear relaxation of the binary problem is optimal. This alignment not only produces decision-aware learning models, but also dramatically reduces training time as it circumvents the need to solve BLPs to compute a loss function with its gradients. Experiments across multiple datasets show that our method exhibits a favorable trade-off between training time and solution quality, particularly with large-scale optimization problems such as vehicle routing, a hard BLP that has yet to benefit from predict-then-optimize methods in the literature due to its difficulty.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 14871
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 9437
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Amos, B., Kolter, J.Z.: Optnet: differentiable optimization as a layer in neural networks. In: International Conference on Machine Learning, pp. 136–145. PMLR (2017)

    Google Scholar 

  2. Ban, G.Y., Rudin, C.: The big data newsvendor: practical insights from machine learning. Oper. Res. 67(1), 90–108 (2019)

    Article  MathSciNet  Google Scholar 

  3. Berthet, Q., Blondel, M., Teboul, O., Cuturi, M., Vert, J.P., Bach, F.: Learning with differentiable perturbed optimizers. arXiv preprint arXiv:2002.08676 (2020)

  4. Dalle, G., Baty, L., Bouvier, L., Parmentier, A.: Learning with combinatorial optimization layers: a probabilistic approach. arXiv preprint arXiv:2207.13513 (2022)

  5. Dantzig, G., Fulkerson, R., Johnson, S.: Solution of a large-scale traveling-salesman problem. J. Oper. Res. Soc. Am. 2(4), 393–410 (1954)

    MathSciNet  Google Scholar 

  6. Elmachtoub, A., Liang, J.C.N., McNellis, R.: Decision trees for decision-making under the predict-then-optimize framework. In: International Conference on Machine Learning, vol. 119, pp. 2858–2867. PMLR (2020)

    Google Scholar 

  7. Elmachtoub, A.N., Grigas, P.: Smart predict, then optimize. Manag. Sci. (2021)

    Google Scholar 

  8. Ferber, A., Wilder, B., Dilkina, B., Tambe, M.: Mipaal: mixed integer program as a layer. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 1504–1511 (2020)

    Google Scholar 

  9. Ferber, A.M., et al.: Surco: learning linear surrogates for combinatorial nonlinear optimization problems. In: International Conference on Machine Learning, pp. 10034–10052. PMLR (2023)

    Google Scholar 

  10. Gurobi Optimization, LLC: Gurobi Optimizer Reference Manual (2021). https://www.gurobi.com

  11. Jeong, J., Jaggi, P., Butler, A., Sanner, S.: An exact symbolic reduction of linear smart predict+ optimize to mixed integer linear programming. In: International Conference on Machine Learning, pp. 10053–10067. PMLR (2022)

    Google Scholar 

  12. Kohl, N., Desrosiers, J., Madsen, O.B., Solomon, M.M., Soumis, F.: 2-path cuts for the vehicle routing problem with time windows. Transp. Sci. 33(1), 101–116 (1999)

    Article  Google Scholar 

  13. Liu, T.Y., et al.: Learning to rank for information retrieval. Found. Trends® Inf. Retr. 3(3), 225–331 (2009)

    Google Scholar 

  14. Mandi, J., Bucarey, V., Tchomba, M.M.K., Guns, T.: Decision-focused learning: through the lens of learning to rank. In: International Conference on Machine Learning, pp. 14935–14947. PMLR (2022)

    Google Scholar 

  15. Mandi, J., Guns, T.: Interior point solving for lp-based prediction+optimisation. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 7272–7282, Curran Associates, Inc. (2020)

    Google Scholar 

  16. Mandi, J., et al.: Decision-focused learning: foundations, state of the art, benchmark and future opportunities. arXiv preprint arXiv:2307.13565 (2023)

  17. Mandi, J., Stuckey, P.J., Guns, T., et al.: Smart predict-and-optimize for hard combinatorial optimization problems. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 1603–1610 (2020). https://doi.org/10.1609/aaai.v34i02.5521

  18. Miller, C.E., Tucker, A.W., Zemlin, R.A.: Integer programming formulation of traveling salesman problems. J. ACM (JACM) 7(4), 326–329 (1960)

    Article  MathSciNet  Google Scholar 

  19. Misra, S., Roald, L., Ng, Y.: Learning for constrained optimization: identifying optimal active constraint sets. INFORMS J. Comput. 34(1), 463–480 (2022)

    Article  MathSciNet  Google Scholar 

  20. Mulamba, M., Mandi, J., Diligenti, M., Lombardi, M., Bucarey, V., Guns, T.: Contrastive losses and solution caching for predict-and-optimize. arXiv preprint arXiv:2011.05354 (2020)

  21. Niepert, M., Minervini, P., Franceschi, L.: Implicit mle: backpropagating through discrete exponential family distributions. Adv. Neural Inf. Process. Syst. 34, 14567–14579 (2021)

    Google Scholar 

  22. Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 32 (2019)

    Google Scholar 

  23. Pogančić, M.V., Paulus, A., Musil, V., Martius, G., Rolinek, M.: Differentiation of blackbox combinatorial solvers. In: International Conference on Learning Representations (2019)

    Google Scholar 

  24. Sadana, U., Chenreddy, A., Delage, E., Forel, A., Frejinger, E., Vidal, T.: A survey of contextual optimization methods for decision making under uncertainty. arXiv preprint arXiv:2306.10374 (2023)

  25. Sahoo, S.S., Paulus, A., Vlastelica, M., Musil, V., Kuleshov, V., Martius, G.: Backpropagation through combinatorial algorithms: identity with projection works. arXiv preprint arXiv:2205.15213 (2022)

  26. Shah, S., Perrault, A., Wilder, B., Tambe, M.: Leaving the nest: going beyond local loss functions for predict-then-optimize. arXiv preprint arXiv:2305.16830 (2023)

  27. Shah, S., Wilder, B., Perrault, A., Tambe, M.: Learning (local) surrogate loss functions for predict-then-optimize problems. arXiv e-prints pp. arXiv–2203 (2022)

    Google Scholar 

  28. Tang, B., Khalil, E.B.: PyEPO: a pytorch-based end-to-end predict-then-optimize library for linear and integer programming. arXiv preprint arXiv:2206.14234 (2022)

  29. Toth, P., Vigo, D.: Vehicle routing: problems, methods, and applications. SIAM (2014)

    Google Scholar 

  30. Virtanen, P., et al.: Scipy 1.0: fundamental algorithms for scientific computing in python. Nat. Methods 17(3), 261–272 (2020)

    Google Scholar 

  31. Wilder, B., Dilkina, B., Tambe, M.: Melding the data-decisions pipeline: decision-focused learning for combinatorial optimization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 1658–1665 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Elias B. Khalil .

Editor information

Editors and Affiliations

A More Experiments

A More Experiments

1.1 A.1 TSP Relaxation

To enhance training efficiency, it has been proposed in [17] that a linear relaxation can be used as a substitute for the original BLP during training. Because the DFJ formulation utilizes constraint generation to handle subtour elimiation constraints, it is challenging to achieve linear relaxation due to the exponential number of constraints. In this study, we used the LP relaxation of the Miller-Tucker-Zemlin (MTZ) formulation [18] of the TSP, and trained SPO+ Rel and PFYL Rel on the same TSP instances with 20 and 50 nodes. Although relaxation methods are more efficient than CaVE in TSP20, they result in higher regret. As the size of the model increases, the relaxation approach also loses its efficiency advantage (Tables 7 and 8).

Table 7. Experimental Results for TSP20 Relaxation
Table 8. Experimental Results for TSP50 Relaxation

1.2 A.2 CVRP Relaxation

Similarly, the CVRP formulation with k-path cuts also struggles to obtain a linear relaxation. Thus, the MTZ formulation can be modified with constraints

$$ u_j - u_i \ge Q (x_{ij} - 1) + q_j \quad \forall i \ne j, i \ne 0, j \ne 0 $$

for subtour elimilation and customer demand, where Q is the capacity and \(q_i\) is the demand of customer i. The SPO+ Rel and PFYL Rel are trained for the same VRP instances with 20 customers. While the linear relaxation approach demonstrates high efficiency in these scenarios, it does not guarantee a low regret compared to CaVE (Table 9).

Table 9. Experimental Results for CVRP20 Relaxation

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tang, B., Khalil, E.B. (2024). CaVE: A Cone-Aligned Approach for Fast Predict-then-optimize with Binary Linear Programs. In: Dilkina, B. (eds) Integration of Constraint Programming, Artificial Intelligence, and Operations Research. CPAIOR 2024. Lecture Notes in Computer Science, vol 14743. Springer, Cham. https://doi.org/10.1007/978-3-031-60599-4_12

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-60599-4_12

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-60601-4

  • Online ISBN: 978-3-031-60599-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics