2020
DOI: 10.1007/s10107-020-01474-5
|View full text |Cite
|
Sign up to set email alerts
|

Strong mixed-integer programming formulations for trained neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 99 publications
(34 citation statements)
references
References 42 publications
0
32
0
Order By: Relevance
“…Since PEREGRiNN is a sound and complete verification algorithm, we restrict our comparison to other sound and complete algorithms. NN verifiers can be grouped into roughly three categories: (i) SMT-based methods, which encode the problem into a Satisfiability Modulo Theory problem [11,18,19] problem as a Mixed Integer Linear Program [3,[5][6][7][8]14,23,29]; (iii) Reachability based methods, which perform layer-by-layer reachability analysis to compute the reachable set [4,13,15,17,30,32,34,35]; and (iv) convex relaxations methods [10,31,33]. In general, (i), (ii) and (iii) suffer from poor scalability.…”
Section: Related Workmentioning
confidence: 99%
“…Since PEREGRiNN is a sound and complete verification algorithm, we restrict our comparison to other sound and complete algorithms. NN verifiers can be grouped into roughly three categories: (i) SMT-based methods, which encode the problem into a Satisfiability Modulo Theory problem [11,18,19] problem as a Mixed Integer Linear Program [3,[5][6][7][8]14,23,29]; (iii) Reachability based methods, which perform layer-by-layer reachability analysis to compute the reachable set [4,13,15,17,30,32,34,35]; and (iv) convex relaxations methods [10,31,33]. In general, (i), (ii) and (iii) suffer from poor scalability.…”
Section: Related Workmentioning
confidence: 99%
“…More complex ML models have also been shown to be MIO-representable, although more effort is required to represent them than simple regression models. Neural networks which use the ReLU activation function can be represented using binary variables and big-M formulations (Amos et al 2016, Grimstad and Andersson 2019, Anderson et al 2020, Chen et al 2020, Spyros 2020, Venzke et al 2020. Where other activation functions are used (Gutierrez-Martinez et al 2011, Lombardi et al 2017, Schweidtmann and Mitsos 2019, the MIO representation of neural networks is still possible, provided the solvers are capable of handling these functions.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Rectified neural networks are known to be continuous piecewise-linear functions with the universal approximation ability [99]. As a result, they are a subject of interest for capturing complex nonlinear physics in MILPs [51,100]. Neural networks are popular because of their ability to model large high-dimensional data-sets; however they also possess drawbacks when considering their application in modeling for use in an MILP.…”
Section: Combined Unit Commitment and Economic Dispatchmentioning
confidence: 99%