ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021
DOI: 10.1109/icassp39728.2021.9414562
|View full text |Cite
|
Sign up to set email alerts
|

L-Red: Efficient Post-Training Detection of Imperceptible Backdoor Attacks Without Access to the Training Set

Abstract: Backdoor attacks (BAs) are an emerging form of adversarial attack typically against deep neural network image classifiers. The attacker aims to have the classifier learn to classify to a target class when test images from one or more source classes contain a backdoor pattern, while maintaining high accuracy on all clean test images. Reverse-Engineering-based Defenses (REDs) against BAs do not require access to the training set but only to an independent clean dataset. Unfortunately, most existing REDs rely on … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 11 publications
(17 citation statements)
references
References 11 publications
0
17
0
Order By: Relevance
“…[48] addressed this issue by estimating a putative BP for each class pair, but this greatly increases the required computation. While [34] and [49] estimate the BA source classes and the target class via a complicated optimization procedure, our method can accurately detect BAs with an arbitrary number of source classes and is computationally efficient, as will be shown in our experiments.…”
Section: Backdoor Defensesmentioning
confidence: 89%
See 2 more Smart Citations
“…[48] addressed this issue by estimating a putative BP for each class pair, but this greatly increases the required computation. While [34] and [49] estimate the BA source classes and the target class via a complicated optimization procedure, our method can accurately detect BAs with an arbitrary number of source classes and is computationally efficient, as will be shown in our experiments.…”
Section: Backdoor Defensesmentioning
confidence: 89%
“…Compared with REDs that may suffer from mismatch between the assumed BP embedding type and the true attack BP type, our optimization problem does not require assuming a BP embedding type. Moreover, REDs' BP estimation using clean samples from all non-target classes has been found experimentally to fail when the majority of these classes are not source classes [49,34]. By contrast, our method can detect BAs with an arbitrary number of source classes, since problem (2) above does not involve any legitimate samples from the domain.…”
Section: Detection Proceduresmentioning
confidence: 98%
See 1 more Smart Citation
“…[40] attempted a robust defense against various attack via data augmentation methods and model fine-tuning. [37] analyzed by reverse engineering triggers through the Lagrangian function.…”
Section: Defense Against Backdoor Attackmentioning
confidence: 99%
“…The main computational cost is induced by the need to determine for each of the K(K − 1) class pairs whether it is involved in a backdoor attack -trigger reverse-engineering is performed for each class pair. Since there is no constraint on the trigger reverse-engineering algorithm used by UMD, the efficiency of UMD can potentially be improved, e.g., by adopting the warm-up strategy by Shen et al (2021) or the weighted-sum strategy by Xiang et al (2021) to accelerate the trigger reverse-engineering process.…”
Section: E4 Computational Cost Of Umdmentioning
confidence: 99%