2020
DOI: 10.1007/978-3-030-58574-7_14
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Robustness on In- and Out-Distribution Improves Explainability

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

1
41
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 37 publications
(42 citation statements)
references
References 21 publications
1
41
0
Order By: Relevance
“…This concept builds on [290]. In a related approach, Augustine et al [291] associate model explainability to its adversarial robustness, demonstrating generative properties of their adversarially robust model similar to [290]. Elliott et al [292] also attempts to bridge the gap between adversarial perturbations and counter-factual explanation of deep models.…”
Section: B the Link Between Attacks And Model Interpretationmentioning
confidence: 99%
“…This concept builds on [290]. In a related approach, Augustine et al [291] associate model explainability to its adversarial robustness, demonstrating generative properties of their adversarially robust model similar to [290]. Elliott et al [292] also attempts to bridge the gap between adversarial perturbations and counter-factual explanation of deep models.…”
Section: B the Link Between Attacks And Model Interpretationmentioning
confidence: 99%
“…Here, we focus on works where robustness to adversarial perturbation is a goal as well. Augustin et al [7] and Sehwag et al [8] consider the problem of combining robust classification and robust OOD detection, like we do. Sehwag et al investigated the robustness of multiple OOD detection methods and found that existing OOD detectors are not robust.…”
mentioning
confidence: 84%
“…In this section, we discuss two common approaches to train models for OOD detection. In the first approach, the goal is to make the model output a uniform distribution when an OOD input is presented [7], [10]. This is implemented using the cross-entropy loss function with the uniform distribution as the true distribution:…”
Section: Training Objectives In Related Workmentioning
confidence: 99%
See 2 more Smart Citations