Interspeech 2019 2019
DOI: 10.21437/interspeech.2019-2818
|View full text |Cite
|
Sign up to set email alerts
|

Acoustic Model Bootstrapping Using Semi-Supervised Learning

Abstract: This work aims at bootstrapping acoustic model training for automatic speech recognition with small amounts of humanlabeled speech data and large amounts of machine-labeled speech data.Semi-supervised learning is investigated to select the machine-transcribed training samples.Two semi-supervised learning methods are proposed: one is the local-global uncertainty based method which introduces both the local uncertainty from the current utterance and the global uncertainty from the whole data pool into the data s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 13 publications
(23 reference statements)
0
2
0
Order By: Relevance
“…We demonstrate that a step-wise distillation approach, introduced in [23] can be effective, although this comes at the cost of more computation at training time. In low data regimes, SSL is an effective technique to reduce annotation costs [14,17,21,5]. For our second task, using knowledge distillation for SSL, we find that to achieve a performance comparable to that of a fully supervised system, the proportion of required supervised data decreases as the amount of total data increases.…”
Section: Introductionmentioning
confidence: 95%
“…We demonstrate that a step-wise distillation approach, introduced in [23] can be effective, although this comes at the cost of more computation at training time. In low data regimes, SSL is an effective technique to reduce annotation costs [14,17,21,5]. For our second task, using knowledge distillation for SSL, we find that to achieve a performance comparable to that of a fully supervised system, the proportion of required supervised data decreases as the amount of total data increases.…”
Section: Introductionmentioning
confidence: 95%
“…• teacher-student learning if non-transcribed data is available in the target domain [19,20,21,22,23,24,25].…”
Section: Introductionmentioning
confidence: 99%