Proceedings of the 27th ACM Symposium on Operating Systems Principles 2019
DOI: 10.1145/3341301.3359639
|View full text |Cite
|
Sign up to set email alerts
|

Privacy accounting and quality control in the sage differentially private ML platform

Abstract: Companies increasingly expose machine learning (ML) models trained over sensitive user data to untrusted domains, such as end-user devices and wide-access model stores. This creates a need to control the data's leakage through these models. We present Sage, a differentially private (DP) ML platform that bounds the cumulative leakage of training data through models. Sage builds upon the rich literature on DP ML algorithms and contributes pragmatic solutions to two of the most pressing systems challenges of glob… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
22
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 25 publications
(23 citation statements)
references
References 39 publications
0
22
0
Order By: Relevance
“…Some of the most prominent demands, which are close in spirit to the requirements of CL, as we need here, are listed as follows: R1: Endless execution R2: Multiple usage of data subsets R3: Capability of changing DP parameters during the execution Satisfying all the R1-R3 together is hard, therefore related papers address only one or two of these requirements. Along this line, two recent DL-based papers of [23], [24] have enabled DP to work on growing databases (dynamic datasets). More explicitly, to address R1 Cummings et al have considered a scheduler to re-execute the DL algorithms whenever the new received data is sufficient [24].…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Some of the most prominent demands, which are close in spirit to the requirements of CL, as we need here, are listed as follows: R1: Endless execution R2: Multiple usage of data subsets R3: Capability of changing DP parameters during the execution Satisfying all the R1-R3 together is hard, therefore related papers address only one or two of these requirements. Along this line, two recent DL-based papers of [23], [24] have enabled DP to work on growing databases (dynamic datasets). More explicitly, to address R1 Cummings et al have considered a scheduler to re-execute the DL algorithms whenever the new received data is sufficient [24].…”
Section: Related Workmentioning
confidence: 99%
“…Note that, since each block of data might be used by different DL algorithms corresponding to the pipelines, calculating the PB spent by the whole pipelines would be challenging. To reach this goal, the authors of [23], have proposed the so-called block composition theorem by which the DL algorithms are executed till the PB consumption of each block 1 does not exceed the predefined GPB. To achieve the desired accuracy, with the aim of re-training the pipelines, either the relevant PB of each pipeline or the number of available samples is doubled.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations