2023
DOI: 10.3390/e25070974
|View full text |Cite
|
Sign up to set email alerts
|

Backdoor Attack against Face Sketch Synthesis

Abstract: Deep neural networks (DNNs) are easily exposed to backdoor threats when training with poisoned training samples. Models using backdoor attack have normal performance for benign samples, and possess poor performance for poisoned samples manipulated with pre-defined trigger patterns. Currently, research on backdoor attacks focuses on image classification and object detection. In this article, we investigated backdoor attacks in facial sketch synthesis, which can be beneficial for many applications, such as anima… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 62 publications
(66 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?