2017 IEEE Symposium on Security and Privacy (SP) 2017
DOI: 10.1109/sp.2017.41
|View full text |Cite
|
Sign up to set email alerts
|

Membership Inference Attacks Against Machine Learning Models

Abstract: Abstract-We quantitatively investigate how machine learning models leak information about the individual data records on which they were trained. We focus on the basic membership inference attack: given a data record and black-box access to a model, determine if the record was in the model's training dataset. To perform membership inference against a target model, we make adversarial use of machine learning and train our own inference model to recognize differences in the target model's predictions on the inpu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

38
2,990
1
8

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 2,684 publications
(3,037 citation statements)
references
References 28 publications
(48 reference statements)
38
2,990
1
8
Order By: Relevance
“…C ) Compute per-example gradients of discriminator loss on fake data Z t and clip them 13 for i ∈ Z t do 14 Compute grad d f ake t ← − ∇ θ d d loss f ake(θ d t , Z i ) 15 grad d f ake t = grad d f ake t /max(1, ||grad d f ake ||2 C ) 16 Compute the overall gradients of discriminator and add Gaussian Noise to them 17 grad d t ← − 1 bs grad d real t + grad d f ake t + N (0, σ 2 C 2 I) 18 Take the gradient Descent step for discriminator 19 θ dt+1 ← − SGD(grads d t , θ dt , lr)) / * Update RDP Accountant * / 20 Accumulate the spent privacy budget using RDP Accountant / * Update the Generator Network * / 21 g loss ← − log(1 − D(G(Z t ))) 22 Compute gradients of generator loss 23 Compute grad g t ← − ∇ θg g loss(θ g t , Z i ) 24 Take the gradient Descent step for generator 25 θ g t+1 ← − ADAM (grad g t , θ g t ) 26 if spent epsilon > OR spent delta > δ then The dataset used used in the evaluation is MNIST handwritten dataset containing 60k training samples and 10k test samples. In the experiments, batch size is set to 600, δ = 10 −5 and learning rate is set by an adapative approach in which the initial learning rate is 0.15, it is decreased to 0.052 in iteration 10K and is fixed on 0.052 for the rest iterations.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…C ) Compute per-example gradients of discriminator loss on fake data Z t and clip them 13 for i ∈ Z t do 14 Compute grad d f ake t ← − ∇ θ d d loss f ake(θ d t , Z i ) 15 grad d f ake t = grad d f ake t /max(1, ||grad d f ake ||2 C ) 16 Compute the overall gradients of discriminator and add Gaussian Noise to them 17 grad d t ← − 1 bs grad d real t + grad d f ake t + N (0, σ 2 C 2 I) 18 Take the gradient Descent step for discriminator 19 θ dt+1 ← − SGD(grads d t , θ dt , lr)) / * Update RDP Accountant * / 20 Accumulate the spent privacy budget using RDP Accountant / * Update the Generator Network * / 21 g loss ← − log(1 − D(G(Z t ))) 22 Compute gradients of generator loss 23 Compute grad g t ← − ∇ θg g loss(θ g t , Z i ) 24 Take the gradient Descent step for generator 25 θ g t+1 ← − ADAM (grad g t , θ g t ) 26 if spent epsilon > OR spent delta > δ then The dataset used used in the evaluation is MNIST handwritten dataset containing 60k training samples and 10k test samples. In the experiments, batch size is set to 600, δ = 10 −5 and learning rate is set by an adapative approach in which the initial learning rate is 0.15, it is decreased to 0.052 in iteration 10K and is fixed on 0.052 for the rest iterations.…”
Section: Resultsmentioning
confidence: 99%
“…Some previous studies have proposed approaches to addressing the problem of preserving privacy in Deep Learning. Shokri et al [22] developed a distributed approach in which multiple parties train a model on their local training set independently. Then, each party selects a set of key parameters, and shares them with the other parties.…”
Section: Related Workmentioning
confidence: 99%
“…Table 4 illustrates the attack results on matrix-factorization-based recommender systems when we weight normal users, where the experimental settings are the same as those in Table 1. Here, "Weighting" means that we weight each normal user and optimize the attack of (30) over the weighted normal users, and the weight of each normal user is computed based on (31). Comparing Tables 1 and 4, we can see that the performance is improved when we consider the weights of different normal users with respect to the target items.…”
Section: Weighting Normal Usersmentioning
confidence: 99%
“…Often, an aggregator is a central entity that also redistributes the merged model parameters to all participants but other topologies have been used as well, e.g., co-locating an aggregator with each participant. However, this approach still poses privacy risks: inference attacks in the learning phase have been proposed by [30]; deriving private information from a trained model has been demonstrated in [37]; and a model inversion attack has been presented in [19].…”
Section: Introductionmentioning
confidence: 99%