Interspeech 2015 2015
DOI: 10.21437/interspeech.2015-696
|View full text |Cite
|
Sign up to set email alerts
|

Recurrent neural network language model adaptation for multi-genre broadcast speech recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
20
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 50 publications
(22 citation statements)
references
References 28 publications
2
20
0
Order By: Relevance
“…The results show a very significant drop in perplexity when using RNNLMs but only a modest improvement in word error rate of 0.7%. This is consistent with the experiments reported on the same BBC data in [35]. The main difference, however is that in [35], instead of LM 1 as background language model, another corpus of 1 billion words was used for language modelling, and different topic models including LDA, were used to classify the text into a set of different genres.…”
Section: Multi-genre Language Modellingsupporting
confidence: 85%
“…The results show a very significant drop in perplexity when using RNNLMs but only a modest improvement in word error rate of 0.7%. This is consistent with the experiments reported on the same BBC data in [35]. The main difference, however is that in [35], instead of LM 1 as background language model, another corpus of 1 billion words was used for language modelling, and different topic models including LDA, were used to classify the text into a set of different genres.…”
Section: Multi-genre Language Modellingsupporting
confidence: 85%
“…More recently, neural adapation approaches have been used to adapt a LM to a target domain based on non-linguistic contextual signals, such as the application at the time of the request [11], or learned topic vectors [12,13]. For example, arXiv:2101.03229v1 [cs.CL] 5 Jan 2021 [12] used topic representations obtained from latent dirichlet allocation to adapt an NLM for genres and shows in a multi-genre broadcast transcription task. Domain-adaptation can also be achieved via shallow-fusion, in which an external (contextually constrained) LM is integrated during beam search [14].…”
Section: Previous Workmentioning
confidence: 99%
“…[17,17,18,19] investigated the use of topic information to build improved n-gram LMs for first-pass decoding. In [20,21], the topic information was modelled in RNNLMs as additional features and used for rescoring. [22] build a session-level LSTM-LM to capture the session-level information.…”
Section: Improving Lm With Long-term Historymentioning
confidence: 99%