2017
DOI: 10.1016/j.csl.2017.05.001
|View full text |Cite
|
Sign up to set email alerts
|

Reversible speaker de-identification using pre-trained transformation functions

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
24
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 32 publications
(25 citation statements)
references
References 19 publications
1
24
0
Order By: Relevance
“…The second approach for speaker de-identification used in this work was first presented in [11], and it consists in using a bunch of FW + AS functions pre-trained on a multi-speaker database. Given…”
Section: Pre-trained Transformation Functionsmentioning
confidence: 99%
See 2 more Smart Citations
“…The second approach for speaker de-identification used in this work was first presented in [11], and it consists in using a bunch of FW + AS functions pre-trained on a multi-speaker database. Given…”
Section: Pre-trained Transformation Functionsmentioning
confidence: 99%
“…A plausible solution to this issue is the use of de-identification, which is a process by which a data custodian alters or removes identifying information from a dataset, making it harder for users of the data to determine the identities of the data subjects [6]. Speaker deidentification is usually carried out by either performing automatic speech recognition (ASR) followed by a text-to-speech (TTS) system [7] or applying voice conversion techniques [8][9][10][11]. The latter approach is more extended since it allows the recovery of the original signal and, in addition, it does not rely on the availability and performance of ASR and TTS modules for a given language.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The phase vocoder and the standard vocal tract length normalization were used to conceal the gender of the speaker [6] showing that for such a de-identification system the preceding gender recognition is necessary. Speaker de-identification gender conversion was investigated also in [7] where the spectral amplitude scaling was combined with the piecewise linear transformation and the linear modification of the fundamental frequency (F 0 ) giving 96.9% de-identification accuracy by the speaker identification in the i -vector space. The pre-calculated voice transformations based on GMM mapping and harmonic plus stochastic models with the target synthetic HMM-based voice were used for successful de-identification in the open set comparable to the closed-set de-identification 87.4% open-set vs 91% closed-set de-identification rate.…”
Section: Introductionmentioning
confidence: 99%
“…This issue can be overcome by means of deidentification, which is a process by which a data custodian alters or removes identifying information from a dataset, making it harder for users of the data to determine the identities of the data subjects [4]. The most extended technique for speaker deidentification consists in applying voice conversion techniques [5,6,7,8] in order to modify the voice characteristics of a speaker in a way that, afterwards, they sound like a different speaker.…”
Section: Introductionmentioning
confidence: 99%