Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security 2017
DOI: 10.1145/3128572.3140444
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Examples Are Not Easily Detected

Abstract: Neural networks are known to be vulnerable to adversarial examples: inputs that are close to natural inputs but classified incorrectly. In order to better understand the space of adversarial examples, we survey ten recent proposals that are designed for detection and compare their efficacy. We show that all can be defeated by constructing new loss functions. We conclude that adversarial examples are significantly harder to detect than previously appreciated, and the properties believed to be intrinsic to adver… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

11
1,277
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 1,264 publications
(1,291 citation statements)
references
References 31 publications
11
1,277
1
Order By: Relevance
“…Besides, some defenses to adversarial attack are claimed not robust enough in last few years [48,49] and other methods came up [50]. In general, adversarial crafting attack is a big, important, and popular topic; we have not given a complete analysis to MalNet for adversarial attack, and in the future we will consider conducting some exploration and detailed analysis with relevant experiments for evaluation.…”
Section: Discussionmentioning
confidence: 99%
“…Besides, some defenses to adversarial attack are claimed not robust enough in last few years [48,49] and other methods came up [50]. In general, adversarial crafting attack is a big, important, and popular topic; we have not given a complete analysis to MalNet for adversarial attack, and in the future we will consider conducting some exploration and detailed analysis with relevant experiments for evaluation.…”
Section: Discussionmentioning
confidence: 99%
“…Various empirical defenses (e.g., [17,27,31]) have been proposed to defend against adversarial examples. However, these defenses were often soon broken by adaptive attacks [1,7]. In response, various certified defenses (e.g., [10,15,32,33,37]) against adversarial examples have been developed.…”
Section: Related Workmentioning
confidence: 99%
“…As an exploratory work and logical consequence of the transferability results, we analyze the impact of considering an ensemble of quantized models in order to filter out adversarial examples with a minimum impact on the natural accuracy. Such an ensemble method, like any other detection-based approach, suffers from a narrow threat model since the defense is useless with an attacker aware of the implementation details of the model in the target device [73]. However, for black-box paradigms, the use of quantized ensemble may have an interesting impact on the transferability when associated to other and complementary defense mechanisms.…”
Section: Resultsmentioning
confidence: 99%