2016 7th IEEE International Conference on Cognitive Infocommunications (CogInfoCom) 2016
DOI: 10.1109/coginfocom.2016.7804538
|View full text |Cite
|
Sign up to set email alerts
|

Adding filled pauses and disfluent events into language models for speech recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…In the further research, we want to focus on eliminating common recognition errors by introducing unsupervised language model adaptation to the current topic and specific speaker speaking style and statistical modeling of most frequent hesitation fillers in spontaneous speech for improving system performance and automatic transcription and annotation of large acoustic corpora of the spoken Slovak language [28]. Also, we are planning to append the fully-annotated data from that corpus to the current training data in order to retrain the present acoustic and language models.…”
Section: Resultsmentioning
confidence: 99%
“…In the further research, we want to focus on eliminating common recognition errors by introducing unsupervised language model adaptation to the current topic and specific speaker speaking style and statistical modeling of most frequent hesitation fillers in spontaneous speech for improving system performance and automatic transcription and annotation of large acoustic corpora of the spoken Slovak language [28]. Also, we are planning to append the fully-annotated data from that corpus to the current training data in order to retrain the present acoustic and language models.…”
Section: Resultsmentioning
confidence: 99%
“…In the same vein, our work focuses on modeling false starts in text. Most similar to our work, [4] introduces disfluencies to clean text and uses the processed text for the language model. They, however, use an existing set of disfluencies to be introduced in the text.…”
Section: Related Workmentioning
confidence: 95%
“…In the context of language modeling, quite a few researchers [3,4,17,18,19,20] have studied disfluencies like false starts, filled pauses and repetition in textual data. Some of these work [3,4] have noted that modeling disfluencies can be beneficial for language modeling. In the same vein, our work focuses on modeling false starts in text.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Generally speaking, human-computer speech communication has been one of the most popular topics in CogInfoCom research area. In paper [31], the authors discuss including filled pauses and disfluent events into the training data for statistical language modeling, in order to improve speech recognition accuracy and robustness in the case of spontaneous speech. In [30], speech analysis has been conducted to verify the speaker authorization and measure the stress level within the air-ground voice communication, to improve voice communication in air traffic management security.…”
Section: Related Workmentioning
confidence: 99%