2019
DOI: 10.48550/arxiv.1906.06316
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Towards Stable and Efficient Training of Verifiably Robust Neural Networks

Abstract: Training neural networks with verifiable robustness guarantees is challenging. Several existing successful approaches utilize relatively tight linear relaxation based bounds of neural network outputs, but they can slow down training by a factor of hundreds and over-regularize the network. Meanwhile, interval bound propagation (IBP) based training is efficient and significantly outperforms linear relaxation based methods on some tasks, yet it suffers from stability issues since the bounds are much looser. In th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
51
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 51 publications
(51 citation statements)
references
References 27 publications
0
51
0
Order By: Relevance
“…We show that although AuditAI considers perturbations in the latent space that allows a large range of semantic variations in the pixel-space, it retains the theoretical guarantees of provable verification analogous to IBP-based adversarial robustness [25,24,37] that consider pixel-perturbations.…”
Section: Theoretical Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…We show that although AuditAI considers perturbations in the latent space that allows a large range of semantic variations in the pixel-space, it retains the theoretical guarantees of provable verification analogous to IBP-based adversarial robustness [25,24,37] that consider pixel-perturbations.…”
Section: Theoretical Resultsmentioning
confidence: 99%
“…Second, our model should be verifiable, where the latent space is learned such that the corresponding S i,in is as large as possible given a specification function F . In this case, AuditAI can be seen as doing certified training [25] to improve the verifiability of the model. Finally, the latent space should be able to perform well on the downstream task through f d .…”
Section: Certified Training Through Latent Representationmentioning
confidence: 99%
See 2 more Smart Citations
“…An upper bound on the adversarial loss is also computed in Raghunathan et al (2018a) by solving instead a semidefinite program. Other more scalable and effective methods based on minimizing an upper bound of the adversarial loss have been introduced Balunovic and Vechev, 2019;Dvijotham et al, 2018a;Zhang et al, 2019).…”
Section: Introductionmentioning
confidence: 99%