2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01617
|View full text |Cite
|
Sign up to set email alerts
|

Black-box Detection of Backdoor Attacks with Limited Information and Data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 49 publications
(27 citation statements)
references
References 19 publications
0
21
0
Order By: Relevance
“…Model manipulations require an adversary to be able to influence the training process/data or even control the model. This is enabled by poisoning attacks [43,77,78] or constituted with query-based access only [24,34,57]; for instance, if models are deployed in embedded systems or on MLaaS platforms. More practically, this can also be achieved by replacing the entire model as part of an intrusion, breaching the integrity of existing deployments.…”
Section: B Model Manipulationmentioning
confidence: 99%
“…Model manipulations require an adversary to be able to influence the training process/data or even control the model. This is enabled by poisoning attacks [43,77,78] or constituted with query-based access only [24,34,57]; for instance, if models are deployed in embedded systems or on MLaaS platforms. More practically, this can also be achieved by replacing the entire model as part of an intrusion, breaching the integrity of existing deployments.…”
Section: B Model Manipulationmentioning
confidence: 99%
“…Moreover, training the shadow networks requires a relatively large number of clean samples and is highly computational. REDs, another family of PT defenses, trial reverse-engineer the BP for each putative target class [42,14,25,9,43,45]. Such reverse-engineering is performed using a small set of clean samples possessed by the defender [42], or using simulated samples obtained by model inversion [3,11].…”
Section: Backdoor Defensesmentioning
confidence: 99%
“…Existing PT defenses typically assume that the defender independently possesses a small set of clean, legitimate samples from every class. These samples may be used: i) to reverse-engineer putative BPs, which are the basis for anomaly detection [42,14,48,25,9,43,45,34,49]; or ii) to train shadow neural networks with and without (known) BAs -based on which a binary "meta-classifier" is trained to predict whether the classifier under inspection is backdoor attacked [18,51,40]. However, these methods assume the BP type (the mechanism for embedding a BP) used by the attacker is known.…”
Section: Introductionmentioning
confidence: 99%
“…These methods [48], [49], [50] detect poisoned images by reversing potential triggers contained in given suspicious DNNs. They have a latent assumption that the triggers should be sample-agnostic and the attack should be targeted.…”
Section: Resistance To Trigger Synthesis Based Detectionsmentioning
confidence: 99%