2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2012
DOI: 10.1109/icassp.2012.6289049
|View full text |Cite
|
Sign up to set email alerts
|

Semi-supervised discriminative language modeling for Turkish ASR

Abstract: We present our work on semi-supervised learning of discriminative language models where the negative examples for sentences in a text corpus are generated using confusion models for Turkish at various granularities, specifically, word, subword, syllable and phone levels. We experiment with different language models and various sampling strategies to select competing hypotheses for training with a variant of the perceptron algorithm. We find that morph-based confusion models with a sample selection strategy aim… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2013
2013
2024
2024

Publication Types

Select...
2
2
2

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 14 publications
(22 reference statements)
0
3
0
Order By: Relevance
“…our claim of rapid adaptability of the system to varying mismatched acoustic and linguistic conditions. The extreme mismatched conditions involved in our experiments supports the possibility of going one step further and training our system on artificially generated data of noisy transformations of phrases as in [35,36,38,[57][58][59]. Thus possibly eliminating the need for an ASR for training purposes.…”
Section: F) Adaptationmentioning
confidence: 73%
See 1 more Smart Citation
“…our claim of rapid adaptability of the system to varying mismatched acoustic and linguistic conditions. The extreme mismatched conditions involved in our experiments supports the possibility of going one step further and training our system on artificially generated data of noisy transformations of phrases as in [35,36,38,[57][58][59]. Thus possibly eliminating the need for an ASR for training purposes.…”
Section: F) Adaptationmentioning
confidence: 73%
“…Further, our work is different from discriminative training of acoustic [33] models and discriminative language models (DLM) [34], which are trained directly to optimize the word error rate using the reference transcripts. DLMs in particular involve optimizing, tuning, the weights of the language model with respect to the reference transcripts and are often utilized in re-ranking n-best ASR hypotheses [34][35][36][37][38]. The main distinction and advantage with our method is the NCPCM can potentially re-introduce learning from past mistakes: improving automatic speech recognition output via noisy-clean unseen or pruned-out phrases.…”
Section: Introductionmentioning
confidence: 99%
“…A comparison of various training methods for DLMs is given in [10]. Recently, semi-supervised discriminative language modeling has also been explored [11] [12] [13]. The work in [4] is the most related to our approach.…”
Section: Precious Workmentioning
confidence: 99%