2021
DOI: 10.48550/arxiv.2103.02683
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Preventing Unauthorized Use of Proprietary Data: Poisoning for Secure Dataset Release

Abstract: Large organizations such as social media companies continually release data, for example user images. At the same time, these organizations leverage their massive corpora of released data to train proprietary models that give them an edge over their competitors. These two behaviors can be in conflict as an organization wants to prevent competitors from using their own data to replicate the performance of their proprietary models. We solve this problem by developing a data poisoning method by which publicly rel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(14 citation statements)
references
References 25 publications
0
14
0
Order By: Relevance
“…Early availability attacks focused on simple settings like logistic regression, and support vector machines [Biggio et al, 2012, Muñoz-González et al, 2017. Recently, heuristics have been leveraged to perform availability attacks on deep networks [Feng et al, 2019, Fowl et al, 2021. In contrast to availability attacks, integrity attacks focus on causing a victim model to misclassify a select set of targets.…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…Early availability attacks focused on simple settings like logistic regression, and support vector machines [Biggio et al, 2012, Muñoz-González et al, 2017. Recently, heuristics have been leveraged to perform availability attacks on deep networks [Feng et al, 2019, Fowl et al, 2021. In contrast to availability attacks, integrity attacks focus on causing a victim model to misclassify a select set of targets.…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…In general, data poisoning attacks perturb training data to intentionally cause some malfunctions of the target model [Biggio and Roli, 2018, Goldblum et al, 2020, Schwarzschild et al, 2021. A common class of poisoning attacks aim to cause test-time error on some given samples [Koh and Liang, 2017, Muñoz-González et al, 2017, Chen et al, 2017, Koh et al, 2018, Shafahi et al, 2018 or on all unseen samples [Biggio et al, 2012, Feng et al, 2019, Liu and Shroff, 2019, Shen et al, 2019, Huang et al, 2021, Yuan and Wu, 2021, Fowl et al, 2021a. The latter attacks are also known as indiscriminate poisoning attacks as they do not have specific target examples [Barreno et al, 2010].…”
Section: Related Workmentioning
confidence: 99%
“…The perturbations are restricted to be small and within in a set ∆. Directly solving Equation ( 1) is intractable for deep neural networks and recent works have designed multiple approximate solutions [Feng et al, 2019, Fowl et al, 2021a, Yuan and Wu, 2021. Feng et al [2019] use multiple rounds of optimization to generate perturbations.…”
Section: The Alternating Optimization Approachmentioning
confidence: 99%
See 2 more Smart Citations