2017
DOI: 10.1007/978-3-319-54472-4_48
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Language Model Adaptation by Data Selection for Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
4

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 18 publications
0
2
0
Order By: Relevance
“…This approach will retain the linguistic regularities encapsulated within original pre-trained NLM, given that embeddings of the rare words are properly modified. Our method can be also viewed as a language model adaptation task [25] where instead of topic or speaking style the vocabulary is adapted to conform with the words used in the target domain.…”
Section: Embedding Matrix Augmentationmentioning
confidence: 99%
“…This approach will retain the linguistic regularities encapsulated within original pre-trained NLM, given that embeddings of the rare words are properly modified. Our method can be also viewed as a language model adaptation task [25] where instead of topic or speaking style the vocabulary is adapted to conform with the words used in the target domain.…”
Section: Embedding Matrix Augmentationmentioning
confidence: 99%
“…This is due to the 'locked-in' phenomenon, also referred to as error propagation, where increasing the probabilities of misrecognized words imposes the ASR system to repeat the same mistakes. To mitigate the 'locked-in' phenomenon, Khassanov et al [107] proposed to use cache data to select relevant sentences from the generic background corpus, and then to use selected data to update the background LM. Although the proposed method avoids the direct usage of cache data, it will increase the latency and introduce additional complexities such as reliable data selection process.…”
Section: )mentioning
confidence: 99%