1990
DOI: 10.1109/34.56193
|View full text |Cite
|
Sign up to set email alerts
|

A cache-based natural language model for speech recognition

Abstract: Speech recognition systems must often decide between competing ways of breaking up the acoustic input into strings of words. Since the possible strings may be acoustically similar, a language model is required; given a word string, the model returns its linguistic probability. This thesis discusses several Markov language models. Subsequently, we present a new kind of language model which • Acknowledgements

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
224
0
7

Year Published

1996
1996
2015
2015

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 370 publications
(233 citation statements)
references
References 8 publications
0
224
0
7
Order By: Relevance
“…However, in this paper, we show that structure reuse is one possible way in which the independence assumption is broken. A simple and principled approach to handling structure re-use is to use adaptation probabilities for probabilistic grammar rules, analogous to cache probabilities used in caching language models (Rosenfeld, Wasserman, Cai, & Zhu, 1999;Kuhn & Mori, 1990), which is what we proposed in this paper.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…However, in this paper, we show that structure reuse is one possible way in which the independence assumption is broken. A simple and principled approach to handling structure re-use is to use adaptation probabilities for probabilistic grammar rules, analogous to cache probabilities used in caching language models (Rosenfeld, Wasserman, Cai, & Zhu, 1999;Kuhn & Mori, 1990), which is what we proposed in this paper.…”
Section: Discussionmentioning
confidence: 99%
“…The importance of comprehension priming at the lexical level has also been noted by the speech recognition community (Kuhn & Mori, 1990), who use socalled caching language models to improve the performance of speech comprehension software. The concept of caching language models is quite simple: a cache of recently seen words is maintained, and the probability of words in the cache is higher than those outside the cache.…”
Section: Adaptationmentioning
confidence: 90%
“…One means of injecting long-range awareness into a language model is by retaining a cache of the most recently seen n-grams which is combined (typically by linear interpolation) with the static model (Jelinek et al, 1991;Kuhn & de Mori, 1990). Another approach, using maximum entropy methods, introduces parameters for trigger pairs of mutually informative words, so that occurrences of certain words in recent context boost the probabilities of the words that they trigger (Lau, Rosenfeld, & Roukos, 1993).…”
Section: Some Doctors Are More Skilled At Doing the Procedures Than Otmentioning
confidence: 99%
“…In this approach, models trained on out-of-domain data can be interpolated with models trained on the last Q sentences that have been processed [10,30,26] and the performance in SMT has been carefully assessed in [47].…”
Section: Adaptationmentioning
confidence: 99%