2019
DOI: 10.48550/arxiv.1911.08040
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Poison as a Cure: Detecting & Neutralizing Variable-Sized Backdoor Attacks in Deep Neural Networks

Abstract: Deep learning models have recently shown to be vulnerable to backdoor poisoning, an insidious attack where the victim model predicts clean images correctly but classifies the same images as the target class when a trigger poison pattern is added. This poison pattern can be embedded in the training dataset by the adversary. Existing defenses are effective under certain conditions such as a small size of the poison pattern, knowledge about the ratio of poisoned training samples or when a validated clean dataset … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 18 publications
(15 citation statements)
references
References 23 publications
0
13
0
Order By: Relevance
“…However, it points out that the outlier ration is fixed to be close to the ratio of corrupted samples in the target class. This requires some knowledge of the poison ratio and target class [142], [143], which turns to be unknown in practice.…”
Section: B Offline Inspectionmentioning
confidence: 99%
“…However, it points out that the outlier ration is fixed to be close to the ratio of corrupted samples in the target class. This requires some knowledge of the poison ratio and target class [142], [143], which turns to be unknown in practice.…”
Section: B Offline Inspectionmentioning
confidence: 99%
“…First, the defense can only be executed at the server side where only local gradients are available. This invalids many backdoor defense methods developed in centralized machine learning, for example, denoising (preprocessing) methods [33], [34], [35], [36], [37], backdoor sample/trigger detection methods [38], [39], [40], [41], [42], [43], robust data augmentations [44], and finetuning methods [44]. Second, the defense method has to be robust to both data poisoning and model poisoning attacks (e.g., Byzantine, backdoor and Sybil attacks).…”
Section: Secure Flmentioning
confidence: 99%
“…Data inspection methods mainly check whether the input data contains triggers through anomaly detection or just remove the abnormal samples during the inference process. Thus, existing data inspection methods for standalone learning [42], [44]- [47], [132] are appliable for well-trained models by collaborate learning systems. Thus, we summarize model inspection defenses as below, especially the defenses for collaborative learning systems.…”
Section: B Backdoor Defensesmentioning
confidence: 99%