ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9054581
|View full text |Cite
|
Sign up to set email alerts
|

Revealing Backdoors, Post-Training, in DNN Classifiers via Novel Inference on Optimized Perturbations Inducing Group Misclassification

Abstract: With the wide deployment of deep neural network (DNN) classifiers, there is great potential for harm from adversarial learning attacks. Recently, a special type of data poisoning (DP) attack, known as a backdoor, was proposed. These attacks do not seek to degrade classification accuracy, but rather to have the classifier learn to classify to a target class whenever the backdoor pattern is present in a test example. Launching backdoor attacks does not require knowledge of the classifier or its training process … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
49
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 23 publications
(51 citation statements)
references
References 16 publications
(48 reference statements)
2
49
0
Order By: Relevance
“…A BA is typically specified by a target class with label t * ∈ C (|C| = K), a set of source classes S * ⊂ C, where t * / ∈ S * , and a backdoor pattern. Effective backdoor patterns in the literature are either human-imperceptible ( [2,5,18,15]) or human-perceptible ( [1,12,13]). Here we focus on the imperceptible case, where the backdoor pattern is embedded into a clean image x ∈ X by…”
Section: Imperceptible Backdoor Attackmentioning
confidence: 99%
See 4 more Smart Citations
“…A BA is typically specified by a target class with label t * ∈ C (|C| = K), a set of source classes S * ⊂ C, where t * / ∈ S * , and a backdoor pattern. Effective backdoor patterns in the literature are either human-imperceptible ( [2,5,18,15]) or human-perceptible ( [1,12,13]). Here we focus on the imperceptible case, where the backdoor pattern is embedded into a clean image x ∈ X by…”
Section: Imperceptible Backdoor Attackmentioning
confidence: 99%
“…REDs are post-training BA defenses without access to the training set, but with access to the trained classifier and an independent, clean dataset [15,12]. REDs typically consist of a backdoor pattern reverse-engineering/estimation step and an anomaly detection step.…”
Section: Reverse-engineering-based Backdoor Defense (Red)mentioning
confidence: 99%
See 3 more Smart Citations