2004 IEEE International Conference on Acoustics, Speech, and Signal Processing
DOI: 10.1109/icassp.2004.1326091
|View full text |Cite
|
Sign up to set email alerts
|

Improving broadcast news transcription by lightly supervised discriminative training

Abstract: In this paper, we present our experiments on lightly supervised discriminative training with large amounts of broadcast news data for which only closed caption transcriptions are available (TDT data). In particular, we use language models biased to the closedcaption transcripts to recognise the audio data, and the recognised transcripts are then used as the training transcriptions for acoustic model training. A range of experiments that use maximum likelihood (ML) training as well as discriminative training ba… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
57
1

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 75 publications
(59 citation statements)
references
References 9 publications
(16 reference statements)
0
57
1
Order By: Relevance
“…Unsupervised training is similar to lightly supervised training except that a general recognition language model is used, rather than the biased language model for lightly supervised training. This reduces the accuracy of the transcriptions [23], but has been successfully applied to a range of tasks [56,110]. However, the gains from unsupervised approaches can decrease dramatically when discriminative training such as MPE is used.…”
Section: Lightly Supervised and Unsupervised Trainingmentioning
confidence: 99%
See 1 more Smart Citation
“…Unsupervised training is similar to lightly supervised training except that a general recognition language model is used, rather than the biased language model for lightly supervised training. This reduces the accuracy of the transcriptions [23], but has been successfully applied to a range of tasks [56,110]. However, the gains from unsupervised approaches can decrease dramatically when discriminative training such as MPE is used.…”
Section: Lightly Supervised and Unsupervised Trainingmentioning
confidence: 99%
“…These transcriptions are known to be error-full and thus not suitable for direct use when training detailed acoustic models. However, a number of lightly supervised training techniques have been developed to overcome this [23,94,128].…”
Section: Lightly Supervised and Unsupervised Trainingmentioning
confidence: 99%
“…The conventional filtering method, however, has a drawback that it significantly reduces the amount of usable training data. Moreover, it is presumed that the unmatched or less confident segments of the data are more useful than the matched segments because the baseline system failed to recognize them and may be improved with additional training [12]. Recent work by Long et al [14] proposed methods to improve the filtering by considering the phone error rate and confidence measures.…”
Section: Introductionmentioning
confidence: 99%
“…In order to increase the training data for an acoustic model, a scheme of lightly supervised training, which does not require faithful transcripts but exploits available verbatim texts, has been explored for broadcast news [10]- [12] Manuscript received February 6, 2015. Manuscript revised March 13, 2015.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation