Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2020
DOI: 10.48550/arxiv.2006.14076
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Convex Relaxation Barrier, Revisited: Tightened Single-Neuron Relaxations for Neural Network Verification

Christian Tjandraatmadja,
Ross Anderson,
Joey Huchette
et al.

Abstract: We improve the effectiveness of propagation-and linear-optimization-based neural network verification algorithms with a new tightened convex relaxation for ReLU neurons. Unlike previous single-neuron relaxations which focus only on the univariate input space of the ReLU, our method considers the multivariate input space of the affine pre-activation function preceding the ReLU. Using results from submodularity and convex geometry, we derive an explicit description of the tightest possible convex relaxation when… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(13 citation statements)
references
References 23 publications
0
13
0
Order By: Relevance
“…MIO has been gaining traction as a tool to solve challenging learning problems. It has been used for example to tackle sparse regression problems (Wilson and Sahinidis 2017, Atamtürk and Gómez 2019, Bertsimas et al 2020b, Hazimeh et al 2020, Xie and Deng 2020, Gómez and Prokopyev 2021, verification of neural networks (Fischetti and Jo 2018, Khalil et al 2018, Tjandraatmadja et al 2020, and sparse principal component analysis (Dey et al 2018, Bertsimas et al 2020a), among others. More importantly in the context of this paper, MIO methods have been proposed to learn optimal decision trees (Bertsimas and Dunn 2017, Verwer and Zhang 2019, Aghaei et al 2019, 2020, Elmachtoub et al 2020, Mišić 2020.…”
Section: Related Workmentioning
confidence: 99%
“…MIO has been gaining traction as a tool to solve challenging learning problems. It has been used for example to tackle sparse regression problems (Wilson and Sahinidis 2017, Atamtürk and Gómez 2019, Bertsimas et al 2020b, Hazimeh et al 2020, Xie and Deng 2020, Gómez and Prokopyev 2021, verification of neural networks (Fischetti and Jo 2018, Khalil et al 2018, Tjandraatmadja et al 2020, and sparse principal component analysis (Dey et al 2018, Bertsimas et al 2020a), among others. More importantly in the context of this paper, MIO methods have been proposed to learn optimal decision trees (Bertsimas and Dunn 2017, Verwer and Zhang 2019, Aghaei et al 2019, 2020, Elmachtoub et al 2020, Mišić 2020.…”
Section: Related Workmentioning
confidence: 99%
“…These bounds tend to be loose unless optimized during training, which typically comes at a significant cost to standard performance. Further work has aimed to tighten these bounds [7,31,33], however these works focus primarily on small convolutional networks and struggle to scale to more typical deep networks. Other work has studied the limits of these convex relaxations on these small networks using vast amounts of CPU-compute [28].…”
Section: Related Workmentioning
confidence: 99%
“…We defer additional analogous results in the smaller MNIST setting to Appendix E.3. CIFAR10 Much work studying verification of deep networks in the CIFAR10 setting [7,31,33], including the more scalable Lagrangian-based methods [2,10], have focused primarily on a CNN with only 6k hidden units from [36]-smaller than the LeNet architecture used for MNIST [21]. Consequently, the LP relaxation for this network can solved exactly with a commercial LP solver such as Gurobi, and recent SDP solvers can produce even tighter bounds [6], rendering any LPbased solutions for this network obsolete.…”
Section: Improved Bounds From Exact Lp Solutionsmentioning
confidence: 99%
“…They leverage MP to optimize a mixed-integer (linear) problem (MIP) over a polyhedral action space using commercially available solvers such as CPLEX and Gurobi. A number of papers show how trained ReLU-based DNNs can be expressed as an MP with (Tjandraatmadja et al 2020;Anderson et al 2020) also providing ideal reformulations that improve computational efficiencies with a solver. (Ryu et al 2019) propose a Q-learning framework to optimize over continuous action spaces using a combination of MP and a DNN actor.…”
Section: Literature Reviewmentioning
confidence: 99%
“…(Ryu et al 2019) propose a Q-learning framework to optimize over continuous action spaces using a combination of MP and a DNN actor. (Delarue, Anderson, and Tjandraatmadja 2020;van Heeswijk and La Poutré 2019;Xu et al 2020) show how to use ReLU-based DNN value functions to optimize combinatorial problems (e.g., vehicle routing) where the immediate rewards are deterministic and the action space is vast. We extend such approaches and results to problems where the immediate reward can be uncertain as is the case with inventory management problems.…”
Section: Literature Reviewmentioning
confidence: 99%