2017
DOI: 10.48550/arxiv.1702.06280
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the (Statistical) Detection of Adversarial Examples

Abstract: Machine Learning (ML) models are applied in a variety of tasks such as network intrusion detection or malware classification. Yet, these models are vulnerable to a class of malicious inputs known as adversarial examples. These are slightly perturbed inputs that are classified incorrectly by the ML model. The mitigation of these adversarial inputs remains an open problem.As a step towards understanding adversarial examples, we show that they are not drawn from the same distribution than the original data, and c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
291
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 203 publications
(312 citation statements)
references
References 20 publications
2
291
0
Order By: Relevance
“…Szegedy et al [18] was the first work to report the vulnerability of DNN to adversarial samples where they introduced imperceptible adversarial perturbations to handwritten digits images and succeeded to fool the DNN model with high confidence. This discovery has prompted a number of studies in the computer vision community, where several attacks and defenses have been proposed [9,12,10]. There are some works [14,22,11] dealing with the transferability of adversarial attacks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Szegedy et al [18] was the first work to report the vulnerability of DNN to adversarial samples where they introduced imperceptible adversarial perturbations to handwritten digits images and succeeded to fool the DNN model with high confidence. This discovery has prompted a number of studies in the computer vision community, where several attacks and defenses have been proposed [9,12,10]. There are some works [14,22,11] dealing with the transferability of adversarial attacks.…”
Section: Related Workmentioning
confidence: 99%
“…The second defense we consider is the Detect & Reject method [10], which involves training our IDSs to detect not only "abnormal" and "normal" traffic, but also a third class called "adversarial". Thus, whenever the IDS decides that a network traffic record is adversarial, it is rejected.…”
Section: Defenses Against the Transferability Of Adversarial Attacksmentioning
confidence: 99%
“…Two-sample hypothesis testing plays a significant role in a variety of scientific applications, such as bioinformatics, social sciences, and image analysis (Fox and Dimmic, 2006;Osborne et al, 2013;Kohout and Pevnỳ, 2017). As we entering the big data era, high-dimensional and large-scale data is becoming prevalent, particulalry in machine learning and deep learning applications, and the attention to the two-sample testing method for large-scale data is also naturally increasing (Sutherland et al, 2016;Grosse et al, 2017;Carlini and Wagner, 2017;Gao et al, 2020).…”
Section: Introduction 1backgroundmentioning
confidence: 99%
“…To achieve this goal, the defender can, for example, use some AA detection method to discard suspicious inputs. One approach to detect AAs [15,44], is to examine the input to the attacked model. Another approach, which we consider in this paper, is to examine the output of the attacked model.…”
Section: Introductionmentioning
confidence: 99%