2019
DOI: 10.48550/arxiv.1903.05821
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Attribution-driven Causal Analysis for Detection of Adversarial Examples

Susmit Jha,
Sunny Raj,
Steven Lawrence Fernandes
et al.

Abstract: Attribution methods have been developed to explain the decision of a machine learning model on a given input. We use the Integrated Gradient method for finding attributions to define the causal neighborhood of an input by incrementally masking high attribution features. We study the robustness of machine learning models on benign and adversarial inputs in this neighborhood. Our study indicates that benign inputs are robust to the masking of high attribution features but adversarial inputs generated by the stat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 25 publications
0
1
0
Order By: Relevance
“…Similarly, since it relies on deep learning architectures, it might be deceived by adversarial attacks. To mitigate these consequences, several methods have been recently proposed to de-bias deep learning models and make them more robust to adversarial examples [59,60,61,62]. A failure of the system might cause dangerous incidents and have severe consequences on people and facilities [63].…”
Section: Broader Impactmentioning
confidence: 99%
“…Similarly, since it relies on deep learning architectures, it might be deceived by adversarial attacks. To mitigate these consequences, several methods have been recently proposed to de-bias deep learning models and make them more robust to adversarial examples [59,60,61,62]. A failure of the system might cause dangerous incidents and have severe consequences on people and facilities [63].…”
Section: Broader Impactmentioning
confidence: 99%