A Divide and Conquer Algorithm for Predict+Optimize with Non-Convex Problems
Ali Ugur Guler, Emir Demirovic, Jeffrey Chan, James Bailey, Christopher Leckie, Peter J. Stuckey
[AAAI-22] Main Track
Abstract:
The predict+optimize problem combines machine learning and
combinatorial optimization by predicting the problem coefficients first and then using these coefficients to solve the optimization problem.
While this problem can be solved in two separate stages,
recent research shows end to end models can achieve better results.
This requires differentiating through a discrete combinatorial function.
Models that use differentiable surrogates are prone to approximation errors, while existing exact models are limited to dynamic programming, or they do not generalize well with scarce data.
In this work we propose a novel
divide and conquer algorithm based on transition points to reason over exact optimization problems
and predict
the coefficients using the optimization loss. Moreover, our model is not limited to dynamic programming problems.
We also introduce a greedy version, which achieves similar
results with less computation.
In comparison with other predict+optimize frameworks, we show our method outperforms existing exact frameworks and can reason over hard combinatorial problems better than surrogate methods.
combinatorial optimization by predicting the problem coefficients first and then using these coefficients to solve the optimization problem.
While this problem can be solved in two separate stages,
recent research shows end to end models can achieve better results.
This requires differentiating through a discrete combinatorial function.
Models that use differentiable surrogates are prone to approximation errors, while existing exact models are limited to dynamic programming, or they do not generalize well with scarce data.
In this work we propose a novel
divide and conquer algorithm based on transition points to reason over exact optimization problems
and predict
the coefficients using the optimization loss. Moreover, our model is not limited to dynamic programming problems.
We also introduce a greedy version, which achieves similar
results with less computation.
In comparison with other predict+optimize frameworks, we show our method outperforms existing exact frameworks and can reason over hard combinatorial problems better than surrogate methods.
Introduction Video
Sessions where this paper appears
-
Poster Session 2
Fri, February 25 12:45 AM - 2:30 AM (+00:00)
Blue 4
-
Poster Session 9
Sun, February 27 8:45 AM - 10:30 AM (+00:00)
Blue 4