2021
DOI: 10.48550/arxiv.2107.01614
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Survey: Leakage and Privacy at Inference Time

Abstract: Leakage of data from publicly available Machine Learning (ML) models is an area of growing significance as commercial and government applications of ML can draw on multiple sources of data, potentially including users' and clients' sensitive data. We provide a comprehensive survey of contemporary advances on several fronts, covering involuntary data leakage which is natural to ML models, potential malevolent leakage which is caused by privacy attacks, and currently available defence mechanisms. We focus on inf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 9 publications
(8 citation statements)
references
References 112 publications
(206 reference statements)
0
6
0
Order By: Relevance
“…For instance, they can be used to train attacker models that learn to identify both demographic features (implicitly present in the data) and blood test features (explicitly present) that highly correlate with certain diseases. It is then possible to use this trained model to re-identify some patients based on their demographic features and possible combination of diseases (Jegorova et al, 2021).…”
Section: Privacy Attacks In Machine Learning and Healthcarementioning
confidence: 99%
“…For instance, they can be used to train attacker models that learn to identify both demographic features (implicitly present in the data) and blood test features (explicitly present) that highly correlate with certain diseases. It is then possible to use this trained model to re-identify some patients based on their demographic features and possible combination of diseases (Jegorova et al, 2021).…”
Section: Privacy Attacks In Machine Learning and Healthcarementioning
confidence: 99%
“…DRL has just started to be leveraged for maintaining privacy by becoming invariant to private features [130,196,197]. Considering that the privacy issues in machine learning have attracted significant attention [198], we believe that there is a new, emerging domain for learning privacy-preserved disentangled representations. As for every new domain, it will be challenging to connect and exploit the existing concepts, such as the differential privacy [199] and the federated learning [200], with the disentanglement paradigm.…”
Section: Opportunities and Open Challengesmentioning
confidence: 99%
“…In Natural Language Understanding (NLU) applications, the input text often contain sensitive personal information, e.g., racial or ethnic origins, religious or philosophical beliefs, etc [13]. Such information can be directly or indirectly used to identify a specific person, leading to potential privacy leakage that impedes privacyconscious users from releasing data to NLU service providers [3,4,28].…”
Section: Introductionmentioning
confidence: 99%