2020
DOI: 10.2478/popets-2021-0012
|View full text |Cite
|
Sign up to set email alerts
|

The Audio Auditor: User-Level Membership Inference in Internet of Things Voice Services

Abstract: With the rapid development of deep learning techniques, the popularity of voice services implemented on various Internet of Things (IoT) devices is ever increasing. In this paper, we examine user-level membership inference in the problem space of voice services, by designing an audio auditor to verify whether a specific user had unwillingly contributed audio used to train an automatic speech recognition (ASR) model under strict black-box access. With user representation of the input audio data and their corres… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
4
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(9 citation statements)
references
References 34 publications
0
4
0
Order By: Relevance
“…In the literature, there are only a few studies on user-level MI attacks. In Miao et al (2021), the authors investigate MI attacks on speech recognition task to infer if any users' data (voice samples) have been used during training. In Song & Shmatikov (2019), the authors propose a user-level MI attack on text generative models.…”
Section: Membership Inferencementioning
confidence: 99%
See 1 more Smart Citation
“…In the literature, there are only a few studies on user-level MI attacks. In Miao et al (2021), the authors investigate MI attacks on speech recognition task to infer if any users' data (voice samples) have been used during training. In Song & Shmatikov (2019), the authors propose a user-level MI attack on text generative models.…”
Section: Membership Inferencementioning
confidence: 99%
“…Baselines: To the best of our knowledge, there is no user-level MI attack on metric embedding learning. The two user-level MI attacks in literature (Song & Shmatikov, 2019;Miao et al, 2021) require generative models where the victim model's output is a word. Hence, there is no trivial way to adopt them for metric embedding scenario.…”
Section: Experimental Settingsmentioning
confidence: 99%
“…Specifically, a service provider that trains an ML model with user data without adequate consent may violate data protection regulations. In this case, MI can be used to assert whether or not a data sample was used during training, protecting both users and service providers [8,9,10].…”
Section: Introductionmentioning
confidence: 99%
“…MI is, thus, an important aspect of trustworthy machine learning, that should be studied in all its facets and for all types of data. However, while MI has been extensively studied in the realms of image and text data [11], the focus on speech data, particularly in what concerns ASR models, remains limited [12,13,9,14,15,10].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation