2018
DOI: 10.48550/arxiv.1802.03043
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

PoTrojan: powerful neural-level trojan designs in deep learning models

Abstract: With the popularity of deep learning (DL), artificial intelligence (AI) has been applied in many areas of human life. Artificial neural network or neural network (NN), the main technique behind DL, has been extensively studied to facilitate computer vision and natural language processing. However, malicious NNs could bring huge threats in the so-called coming AI era. In this paper, for the first time in the literature, we propose a novel approach to design and insert powerful neuron-level trojans or PoTrojan i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
28
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 26 publications
(28 citation statements)
references
References 13 publications
(12 reference statements)
0
28
0
Order By: Relevance
“…For the first attack strategy, the attacker is assumed to have the perfect knowledge of the target DNN model. In this way, he can directly insert the neuron-level backdoors into the target DNN to modify the structure [16], or maximize the activation of a specific neuron to construct the backdoor [11]. Besides, the attacker can also add the well-designed perturbations into the weight of a specific layer of the target DNN model to embed the backdoor [17], [18], or flip the bits of weight values to inject the backdoor [9].…”
Section: Related Workmentioning
confidence: 99%
“…For the first attack strategy, the attacker is assumed to have the perfect knowledge of the target DNN model. In this way, he can directly insert the neuron-level backdoors into the target DNN to modify the structure [16], or maximize the activation of a specific neuron to construct the backdoor [11]. Besides, the attacker can also add the well-designed perturbations into the weight of a specific layer of the target DNN model to embed the backdoor [17], [18], or flip the bits of weight values to inject the backdoor [9].…”
Section: Related Workmentioning
confidence: 99%
“…These attacks modify a machine learning model through some algorithmic procedure to respond to a specific trigger in the model's input, which, if present, will cause the model to infer a preprogrammed response that could have unknown and potentially malicious consequences in a deployed setting. A trojan attack can be implemented by manipulating both the training data and its associated labels (Gu, Dolan-Gavitt, and Garg 2017), directly altering a model's structure (Zou et al 2018), or adding training data that have correct labels, but are specially-crafted to still produce the trojan behavior (Turner, Tsipras, and Madry 2018). Here, we define a trigger as a model-recognizable characteristic of the input data that is used by an attacker to insert a trojan, and a trojan to be the alternate behavior of the model when exposed to the trigger, as desired by the attacker.…”
Section: Introductionmentioning
confidence: 99%
“…In such scenarios, most of these HT attacks are not applicable due to partitioning of CNN among difference RC devices for horizontal collaboration. Moreover, most of the approaches adopted in state-ofthe-art hardware/firmware Trojan attacks on hardware accelerator based CNN inference is focused on the deployment of CNN on a single FPGA with access to the complete CNN pipeline [12], [15], [17]. Table 1 summarizes these differences in the state-of-the-art HT insertion techniques.…”
Section: Introductionmentioning
confidence: 99%