2023
DOI: 10.48550/arxiv.2302.01428
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dataset Distillation Fixes Dataset Reconstruction Attacks

Abstract: Modern deep learning requires large volumes of data, which could contain sensitive or private information which cannot be leaked. Recent work has shown for homogeneous neural networks a large portion of this training data could be reconstructed with only access to the trained network parameters. While the attack was shown to work empirically, there exists little formal understanding of its effectiveness regime, and ways to defend against it. In this work, we first build a stronger version of the dataset recons… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 17 publications
(20 reference statements)
0
1
0
Order By: Relevance
“…This type of problem arises in many deep learning fields such as hyperparameters optimization (Domke, 2012;MacKay et al, 2019;Maclaurin et al, 2015), meta-learning (Finn et al, 2017;Rajeswaran et al, 2019), and adversarial training Madry et al, 2017;Szegedy et al, 2013) as well as safety and verification methods (Gruenbacher et al, 2022;Grunbacher et al, 2021;Xiao et al, 2022). Similarly, dataset distillation can also be framed as a bilevel optimization problem, with θ the set of network parameters, and ψ our distilled dataset parameters, given by the coreset images and labels (Loo et al, 2023;Nguyen et al, 2021a;Wang et al, 2018;Zhou et al, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…This type of problem arises in many deep learning fields such as hyperparameters optimization (Domke, 2012;MacKay et al, 2019;Maclaurin et al, 2015), meta-learning (Finn et al, 2017;Rajeswaran et al, 2019), and adversarial training Madry et al, 2017;Szegedy et al, 2013) as well as safety and verification methods (Gruenbacher et al, 2022;Grunbacher et al, 2021;Xiao et al, 2022). Similarly, dataset distillation can also be framed as a bilevel optimization problem, with θ the set of network parameters, and ψ our distilled dataset parameters, given by the coreset images and labels (Loo et al, 2023;Nguyen et al, 2021a;Wang et al, 2018;Zhou et al, 2022).…”
Section: Introductionmentioning
confidence: 99%