2017
DOI: 10.48550/arxiv.1708.06733
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain

Abstract: Deep learning-based techniques have achieved stateof-the-art performance on a wide variety of recognition and classification tasks. However, these networks are typically computationally expensive to train, requiring weeks of computation on many GPUs; as a result, many users outsource the training procedure to the cloud or rely on pre-trained models that are then fine-tuned for a specific task. In this paper we show that outsourced training introduces new security risks: an adversary can create a maliciously tr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
907
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 435 publications
(988 citation statements)
references
References 22 publications
2
907
0
Order By: Relevance
“…: Backdoor injection is an emerging attack that leaves backdoors into neural networks during the training process and tricks the trained model to conduct specific behaviors as the backdoor is triggered. In general, different attack methods specify different trigger patterns, which can be one single pixel [17], a tiny patch [9] or human imperceptible noises [18], [19]. This paper is proposed to defend against all kinds of attacks mentioned above.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…: Backdoor injection is an emerging attack that leaves backdoors into neural networks during the training process and tricks the trained model to conduct specific behaviors as the backdoor is triggered. In general, different attack methods specify different trigger patterns, which can be one single pixel [17], a tiny patch [9] or human imperceptible noises [18], [19]. This paper is proposed to defend against all kinds of attacks mentioned above.…”
Section: Related Workmentioning
confidence: 99%
“…Given the recovered trigger patterns, the next step of BAERASER is to erase them through machine unlearning (line [17][18][19]. The basic principle of trigger pattern unlearning is derived from the following observation about gradient descent based neural network learning.…”
Section: B Trigger Pattern Unlearningmentioning
confidence: 99%
See 1 more Smart Citation
“…Therefore, it is challenging to craft adversarial example patch to be robust in diverse real-world scenes. Distinct from the adversarial example attack, recently, there is a new backdoor attack being revealed, with nearly all studies are on classification tasks, especially image classifications [5]- [7]. A backdoored model behaves normally given normal inputs containing the attacker secretly-chosen trigger, but misbehaves as the attacker intends once the trigger is presented in the input.…”
Section: Introductionmentioning
confidence: 99%
“…However, this common practice raises a serious concern that the labeled data from the third parties can be backdoor attacked. Such an operation enables f to perform well on normal samples while behaving badly on samples with specifically designed patterns, leading to serious concerns to DNN (Gu et al, 2017;Li et al, 2020b).…”
Section: Introductionmentioning
confidence: 99%