Proceedings of the 35th Annual Computer Security Applications Conference 2019
DOI: 10.1145/3359789.3359790
|View full text |Cite
|
Sign up to set email alerts
|

Strip

Abstract: A recent trojan attack on deep neural network (DNN) models is one insidious variant of data poisoning attacks. Trojan attacks exploit an effective backdoor created in a DNN model by leveraging the difficulty in interpretability of the learned model to misclassify any inputs signed with the attacker's chosen trojan trigger. Since the trojan trigger is a secret guarded and exploited by the attacker, detecting such trojan inputs is a challenge, especially at run-time when models are in active operation. This work… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 344 publications
(11 citation statements)
references
References 33 publications
0
11
0
Order By: Relevance
“…Based on this requirement, we determinedε by computing the initial logit and then setting a satisfying logit threshold. By forward propagation on the original network, we computed the logit of the random sample X 0 to be Z [4] = −4.8510. In this experiment, we setε = 1, which models a tweaked logit Z [4] ≥ −3.8510.…”
Section: Configuring the Parameter Spacementioning
confidence: 99%
See 2 more Smart Citations
“…Based on this requirement, we determinedε by computing the initial logit and then setting a satisfying logit threshold. By forward propagation on the original network, we computed the logit of the random sample X 0 to be Z [4] = −4.8510. In this experiment, we setε = 1, which models a tweaked logit Z [4] ≥ −3.8510.…”
Section: Configuring the Parameter Spacementioning
confidence: 99%
“…By forward propagation on the original network, we computed the logit of the random sample X 0 to be Z [4] = −4.8510. In this experiment, we setε = 1, which models a tweaked logit Z [4] ≥ −3.8510. The corresponding modeled sensitivity under this setting is β ≥ 0.0131, which satisfies the detection threshold.…”
Section: Configuring the Parameter Spacementioning
confidence: 99%
See 1 more Smart Citation
“…After that, several countermeasures were proposed. Most can be classified as follows: detecting a backdoor model [2,4,10], removing or disabling backdoor neurons from the backdoor model [2,8,12], or removing poison data from a poison training dataset [1]. Liu et al [8] proposed a fine-pruning countermeasure for removing backdoor neurons from a backdoor model.…”
Section: Related Workmentioning
confidence: 99%
“…Countermeasures have been proposed for backdoor attacks. Some have used approaches such as detecting backdoor models [2,4,10], removing or disabling backdoors from backdoor models [2,8,12], and removing poison data from poison training datasets [1].…”
Section: Introductionmentioning
confidence: 99%