2021 IEEE 41st International Conference on Distributed Computing Systems (ICDCS) 2021
DOI: 10.1109/icdcs51616.2021.00086
|View full text |Cite
|
Sign up to set email alerts
|

BaFFLe: Backdoor Detection via Feedback-based Federated Learning

Abstract: Recent studies have shown that federated learning (FL) is vulnerable to poisoning attacks which aim at injecting a backdoor into the global model. These attacks are effective, even when performed by a single client, and undetectable by most existing defensive techniques. In this paper, we propose a novel defense, dubbed BAFFLE-Backdoor detection via Feedback-based Federated Learning-to secure FL against backdoor attacks. The core idea behind BAFFLE is to leverage data of multiple clients not only for training … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
71
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 87 publications
(71 citation statements)
references
References 14 publications
0
71
0
Order By: Relevance
“…An adversary can compromise a subset of the clients and use them to inject a backdoor into the aggregated model. In the examples above the adversary's goal would be to cause the aggregated model to classify malware network traffic patterns as benign to avoid detection by the NIDS, or in the case of NLP to manipulate the text prediction model to propose specific brand names to inconspicuously advertise them 1 . Recently, various attack strategies for targeted poisoning, so-called backdoor attacks, have been proposed utilizing compromised clients to submit poisoned model updates [2], [27], [34], [41], [38].…”
Section: Introductionmentioning
confidence: 99%
“…An adversary can compromise a subset of the clients and use them to inject a backdoor into the aggregated model. In the examples above the adversary's goal would be to cause the aggregated model to classify malware network traffic patterns as benign to avoid detection by the NIDS, or in the case of NLP to manipulate the text prediction model to propose specific brand names to inconspicuously advertise them 1 . Recently, various attack strategies for targeted poisoning, so-called backdoor attacks, have been proposed utilizing compromised clients to submit poisoned model updates [2], [27], [34], [41], [38].…”
Section: Introductionmentioning
confidence: 99%
“…However, only methods with a proper selection of updates to be evaluated can be integrated with our scheme straightforwardly. For example, BaFFLe [58] avoids backdoor attacks by validating the new global model to be updated, which does not leak any information of clients' models to the server, and thus can be adopted for integration. In contrast, defense methods such as [4], [59]- [61] rely on the evaluation of clients' locally trained models, which means that the server must know clients' models.…”
Section: Further Discussionmentioning
confidence: 99%
“…Although it was a promising proposal, the main problem is that in the presence of a non-IID distribution of data between clients it could fail to identify clusters. In Andreina et al [61], they experiment with different anomaly detection mechanisms and combine the results with adaptive clipping and noise. Along the same lines, in Sattler et al [97] the authors propose to divide the model updates into clusters according to the cosine distance and Preuveneers et al [98] proposed an incremental defence based on unsupervised deep learning anomaly detection system integrated in a blockchain process.…”
Section: Anomaly Detectionmentioning
confidence: 99%
“…The attacks are carried out continuously during the training process, either during all the learning rounds or a portion of them. They are more elaborate as the attackers have to become part of the aggregation in several rounds, but this kind of attack can be more effective and stealthy [61].…”
Section: Taxonomy According To the Frequencymentioning
confidence: 99%