Interspeech 2020 2020
DOI: 10.21437/interspeech.2020-1791
|View full text |Cite
|
Sign up to set email alerts
|

A Federated Approach in Training Acoustic Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
41
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 38 publications
(41 citation statements)
references
References 3 publications
0
41
0
Order By: Relevance
“…In [2,7], federated learning was applied to improve a general shared acoustic model with the goal of privacy preservation, but no speaker adaptation was targeted. Federated learning was also experimented in [4] to speed up the training process and improve the shared general acoustic model performance.…”
Section: Related Workmentioning
confidence: 99%
“…In [2,7], federated learning was applied to improve a general shared acoustic model with the goal of privacy preservation, but no speaker adaptation was targeted. Federated learning was also experimented in [4] to speed up the training process and improve the shared general acoustic model performance.…”
Section: Related Workmentioning
confidence: 99%
“…The process restarts and loops until convergence or after a fixed number of rounds. The utility and training efficiency of the FL AMs have been successfully studied in recent works [1][2][3][4][5][6], and these topics are beyond the scope of the current paper. Alternatively, we focus on the privacy aspect of this framework.…”
Section: Federated Learning For Asr Acoustic Modelsmentioning
confidence: 99%
“…Federated learning (FL) for automatic speech recognition (ASR) has recently become an active area of research [1][2][3][4][5][6]. To preserve the privacy of the users' data in the FL framework, the model is updated in a distributed fashion instead of communicating the data directly from clients to a server.…”
Section: Introductionmentioning
confidence: 99%
“…For the LS task, we used byte-pair encoding (BPE) [38] to create 16,000 subword units. The optimizer settings are the same as described in [18]. The LS corpus contains approximately 1000 hours of read speech for training.…”
Section: Accent Adaptation Taskmentioning
confidence: 99%