2013
DOI: 10.1007/s10994-013-5396-x
|View full text |Cite
|
Sign up to set email alerts
|

Differential privacy based on importance weighting

Abstract: This paper analyzes a novel method for publishing data while still protecting privacy. The method is based on computing weights that make an existing dataset, for which there are no confidentiality issues, analogous to the dataset that must be kept private. The existing dataset may be genuine but public already, or it may be synthetic. The weights are importance sampling weights, but to protect privacy, they are regularized and have noise added. The weights allow statistical queries to be answered approximatel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
20
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 27 publications
(20 citation statements)
references
References 18 publications
0
20
0
Order By: Relevance
“…In the Nissim's paper, the conception of local sensitivity is introduced, which can reduce the noise added to the real query answers. And there are papers focusing on privacy protection such as [3] and there are lots of papers focusing on privacypreserving data minting such as [1], [4], [5], [8], [10], [11]. All of these papers are based on noise drawn from Laplace distribution with mean 0.…”
Section: Pinq(privacy Integrated Queriesmentioning
confidence: 99%
“…In the Nissim's paper, the conception of local sensitivity is introduced, which can reduce the noise added to the real query answers. And there are papers focusing on privacy protection such as [3] and there are lots of papers focusing on privacypreserving data minting such as [1], [4], [5], [8], [10], [11]. All of these papers are based on noise drawn from Laplace distribution with mean 0.…”
Section: Pinq(privacy Integrated Queriesmentioning
confidence: 99%
“…In section 3.1, the supervised learning module and the feature extraction module of the proposed framework were evaluated using the subspace (ɑ3, ɑ4, ɑ11) of the NSL-KDD data set with binary classes (9,10) only. In a new experiment, the same subspace is again considered; however, the other classes (0,1), (0,5), (1,2), (1,9), (3,5), (3,9), and (6,8) are also studied. In addition, three other subspaces, (ɑ3, ɑ4, ɑ5), (ɑ3, ɑ4, ɑ7), and (ɑ4, ɑ7, ɑ10), are also included in the experiment to study the performance of the proposed analytical framework with DPLR.…”
Section: Multiple Subspace Analysismentioning
confidence: 99%
“…[6]. Since its introduction, a significant amount of studies have been conducted using this model for achieving privacy strength and prediction/classification accuracy [8,9,12,16]. This is a parametric approach and the selection of its privacy parameter ϵ is a challenging problem.…”
Section: Introductionmentioning
confidence: 99%
“…Assuming there is labeled public data available, the importance weighting approach of Ji and Elkan (2013) can be used for efficient differentially private data publishing. Ji and Elkan (2013) report that the method can reach accurate results already with a small privacy budget, but their example has a much lower dimensionality than any genomic dataset and it is unclear how the method would scale to genomic data.…”
Section: Introductionmentioning
confidence: 99%