Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security 2021
DOI: 10.1145/3460120.3484533
|View full text |Cite
|
Sign up to set email alerts
|

Honest-but-Curious Nets: Sensitive Attributes of Private Inputs Can Be Secretly Coded into the Classifiers' Outputs

Abstract: It is known that deep neural networks, trained for the classification of a non-sensitive target attribute, can reveal sensitive attributes of their input data; through features of different granularity extracted by the classifier. We, taking a step forward, show that deep classifiers can be trained to secretly encode a sensitive attribute of users' input data, at inference time, into the classifier's outputs for the target attribute. An attack that works even if users have a white-box view of the classifier, a… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(10 citation statements)
references
References 59 publications
1
9
0
Order By: Relevance
“…The attackers can then obtain the node embedding matrix from the data holder through the rouge provider. This attack scenario is in line with the malicious machine learning provider scenario discussed by Song et al [65] and Malekzadeh et al [47].…”
Section: Attack Scenariossupporting
confidence: 80%
See 1 more Smart Citation
“…The attackers can then obtain the node embedding matrix from the data holder through the rouge provider. This attack scenario is in line with the malicious machine learning provider scenario discussed by Song et al [65] and Malekzadeh et al [47].…”
Section: Attack Scenariossupporting
confidence: 80%
“…They cannot interact with the node embedding models since such pipelines usually operate in one direction. For instance, the data holder may have integrated with the malicious machine learning solution providers (i.e., MLaaS providers) from the AWS Marketplace [47,65], or the data holder is part of a vertical federated learning environment in an enterprise [71]. In both cases, the node embeddings are part of the learning process and can be obtained by either the malicious MLaaS providers [47,65] or the insiders [71] in the pipeline.…”
Section: Introductionmentioning
confidence: 99%
“…Attribute inference attack is closer to the problem of data amputation where given non-sensitive attributes, the goal is the predict the sensitive attribute [11,14,15,41]. Malekzadeh et al [29] leverage an informational theoretic view to infer the sensitive attributes from output predictions. However, their setting is different where they consider a malicious model designer who injects sensitive attribute to be inferred later after deployment.…”
Section: Related Workmentioning
confidence: 99%
“…Furthermore, given black-box access to a language model's pre-train and finetune stages, Zanella-Béguelin et al (2020) showed that sensitive sequences of the fine-tuning dataset can be extracted. For the distributed client-server setup, Malekzadeh et al (2021) considered the sensitive attribute leakage from the server side with honest-but-curious (HBC) classifiers.…”
Section: Related Workmentioning
confidence: 99%