2020
DOI: 10.48550/arxiv.2006.14026
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Subpopulation Data Poisoning Attacks

Abstract: Machine learning (ML) systems are deployed in critical settings, but they might fail in unexpected ways, impacting the accuracy of their predictions. Poisoning attacks against ML induce adversarial modification of data used by an ML algorithm to selectively change the output of the ML algorithm when it is deployed. In this work, we introduce a novel data poisoning attack called a subpopulation attack, which is particularly relevant when datasets are large and diverse. We design a modular framework for subpopul… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…Conventionally, a poisoning attack is to degrade the model overall inference accuracy for clean samples of its primary task [44]. Poisoning attack is also often called availability attack [45], [46] in a sense that such attack results in lower accuracy of the model, akin to a denial of service attack. In contrast, though a backdoor attack can be realized through data poisoning, a backdoor attack retains the inference accuracy for benign samples of its primary task and only misbehaves in the presence of the secret trigger stealthily.…”
Section: Data Poisoning Attackmentioning
confidence: 99%
See 1 more Smart Citation
“…Conventionally, a poisoning attack is to degrade the model overall inference accuracy for clean samples of its primary task [44]. Poisoning attack is also often called availability attack [45], [46] in a sense that such attack results in lower accuracy of the model, akin to a denial of service attack. In contrast, though a backdoor attack can be realized through data poisoning, a backdoor attack retains the inference accuracy for benign samples of its primary task and only misbehaves in the presence of the secret trigger stealthily.…”
Section: Data Poisoning Attackmentioning
confidence: 99%
“…4) Data Collection: Data collection is usually error-prone and susceptible to untrusted sources [45]. If a user collects training data from multiple sources, then data poisoning attacks become a more realistic threat.…”
Section: Introductionmentioning
confidence: 99%
“…To perform the attack, the adversary only requires access to the labels of the training dataset, however, to optimize the attack, it is often assumed that the adversary has access to the learner's loss function. To fully optimize this attack, the adversary would need either the learning model's parameters and read access to the samples in the training dataset or an auxiliary dataset that follows the same distribution as the training dataset [82].…”
Section: Targeted Attacksmentioning
confidence: 99%
“…To date, there does not seem to be any effective defenses to fight against targeted data poisoning attacks. According to [82], some data poisoning attacks are even impossible to defend against. Our proposed defensive strategies utilize deep learning techniques that may pave the way to the development of more generic defensive mechanisms for various targeted attacks.…”
Section: Targeted Attacksmentioning
confidence: 99%
See 1 more Smart Citation