Interspeech 2019 2019
DOI: 10.21437/interspeech.2019-2678
|View full text |Cite
|
Sign up to set email alerts
|

Unbiased Semi-Supervised LF-MMI Training Using Dropout

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 14 publications
0
7
0
Order By: Relevance
“…Based on this observation, we devised a novel framework which uses Dropout at the test time to sample from the posterior predictive distribution of word-sequences to produce unbiased supervision for semi-supervised training. Results on monolingual experiments show that the proposed approach can further improve the performance over the state-of-the-art method [Tong et al, 2019b].…”
Section: Main Contributionsmentioning
confidence: 99%
See 1 more Smart Citation
“…Based on this observation, we devised a novel framework which uses Dropout at the test time to sample from the posterior predictive distribution of word-sequences to produce unbiased supervision for semi-supervised training. Results on monolingual experiments show that the proposed approach can further improve the performance over the state-of-the-art method [Tong et al, 2019b].…”
Section: Main Contributionsmentioning
confidence: 99%
“…Therefore, we present a novel approach for semi-supervised training in this chapter, which can address the data scarcity problem from a different aspect. The work in this chapter was published as Tong et al [2019b].…”
Section: Semi-supervised Training Using Dropoutmentioning
confidence: 99%
“…A typical approach to exploit unsupervised data for automatic speech recognition (ASR) is to train a seed model using supervised data and use the seed model to automatically transcribe the unsupervised data [1,2,3,4,5,6,7,8,9]. This approach is referred to as semi-supervised learning.…”
Section: Introductionmentioning
confidence: 99%
“…In the aforementioned approaches, the entire unlabelled data is decoded only once with a seed model trained using manually labeled data. In [14,15], multiple systems (or outputs) are used to obtain better labels. In [16], interleaved training by continuously updating the model used to generate labels was shown to be effective.…”
Section: Introductionmentioning
confidence: 99%