2022
DOI: 10.48550/arxiv.2202.07183
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Survey of Neural Trojan Attacks and Defenses in Deep Learning

Abstract: Artificial Intelligence (AI) relies heavily on deep learning -a technology that is becoming increasingly popular in real-life applications of AI, even in the safety-critical and high-risk domains. However, it is recently discovered that deep learning can be manipulated by embedding Trojans inside it. Unfortunately, pragmatic solutions to circumvent the computational requirements of deep learning, e.g. outsourcing model training or data annotation to third parties, further add to model susceptibility to the Tro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 86 publications
0
1
0
Order By: Relevance
“…( 4) Adversarial examples can be used as interpretability tools [43], [67], [117], [241]. (5) Finally, adversarial trojan detection methods can also be used as interpretability/debugging tools [90], [98], [156], [252], [253]. 1 The works referenced in this paragraph are not limited only to inner interpretability methods.…”
Section: Discussionmentioning
confidence: 99%
“…( 4) Adversarial examples can be used as interpretability tools [43], [67], [117], [241]. (5) Finally, adversarial trojan detection methods can also be used as interpretability/debugging tools [90], [98], [156], [252], [253]. 1 The works referenced in this paragraph are not limited only to inner interpretability methods.…”
Section: Discussionmentioning
confidence: 99%
“…The authors suggested a defense mechanism that works by fine-tuning the model on a variety of clean datasets, and they showed that it works on numerous benchmark datasets. Wang et al [81] examined the field of neural trojan attacks. The authors provide an overview of current attack techniques and defense tactics and discuss the significance of creating reliable models to prevent trojan attacks.…”
Section: Susceptibility Of Deep Learning Systems To Backdoor Attacks ...mentioning
confidence: 99%
“…Backdoor (a.k.a. Trojan) attacks manipulate visual models by forcing them to misbehave when exposed to a 'trigger' in the input (Wang, Hassan, and Akhtar 2022). These attacks are stealthy because the model behaves normally for clean inputs, and the model user is unaware of the trigger pattern.…”
Section: Backdoor Detectionmentioning
confidence: 99%
“…In Fig. 4, we show the trigger patterns used in our experiments, which are chosen at random based on the literature (Wang, Hassan, and Akhtar 2022). We apply the proposed input-agnostic saliency mapping to the compromised models.…”
Section: Backdoor Detectionmentioning
confidence: 99%
See 1 more Smart Citation