2020
DOI: 10.1609/aaai.v34i02.5509
|View full text |Cite
|
Sign up to set email alerts
|

MIPaaL: Mixed Integer Program as a Layer

Abstract: Machine learning components commonly appear in larger decision-making pipelines; however, the model training process typically focuses only on a loss that measures average accuracy between predicted values and ground truth values. Decision-focused learning explicitly integrates the downstream decision problem when training the predictive model, in order to optimize the quality of decisions induced by the predictions. It has been successfully applied to several limited combinatorial problem classes, such as tho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 66 publications
(58 citation statements)
references
References 18 publications
0
55
0
Order By: Relevance
“…In case of problems with weaker relaxations, one could consider adding cutting planes prior to solving (Ferber et al 2019). Moreover, further improvements could be achieved by exploiting the fact that all previously computed solutions are valid candidates.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In case of problems with weaker relaxations, one could consider adding cutting planes prior to solving (Ferber et al 2019). Moreover, further improvements could be achieved by exploiting the fact that all previously computed solutions are valid candidates.…”
Section: Discussionmentioning
confidence: 99%
“…In case of weak MIP relaxations, one can also use a cutting plane algorithm in the root node and use the resulting tighter relaxation thereof (Ferber et al 2019). Other weaker oracles could also be used, for example setting a time-limit on an any-time solver and using the best solution found, or a node-limit on search algorithms.…”
Section: Combinatorial Problems and Scaling Upmentioning
confidence: 99%
“…Following this idea, existing work developed implicit layers of argmin in neural network, including OptNet [13] for quadratic programs (QP) problems and CVXPY [14] for more general convex optimization problems. Further with linear relaxation and QP regularization, Wilder et al derived an end-to-end framework for combinatorial programs [9], which accelerates the computation by leverage the low-rank properties of decision vectors [3], and is further extended to mixed integer linear programs in MIPaaL [27]. Besides, for the relaxed LP problems, instead of differentiating KKT conditions, IntOpt [28] proposes an interior point based approach which computes gradients by differentiating homogeneous self-dual formulation.…”
Section: Related Workmentioning
confidence: 99%
“…A series of works about differentiating through CO problems , Elmachtoub and Grigas, 2020, Ferber et al, 2020 relax ILPs by adding L 1 , L 2 or log-barrier regularization terms and differentiate through the KKT conditions deriving from the application of the cutting plane or the interior-point methods. These approaches are conceptually linked to techniques for differentiating through smooth programs , Donti et al, 2017, Agrawal et al, 2019, Chen et al, 2020, Domke, 2012, Franceschi et al, 2018 that arise not only in modelling but also in hyperparameter optimization and meta-learning.…”
Section: Related Workmentioning
confidence: 99%