Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96
DOI: 10.1109/icslp.1996.607136
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation of a language model using a clustered model backoff

Abstract: In this paper, we describe and evaluate a language model using word classes automatically generated from a word clustering algorithm. Class based language models have been shown to be effective for rapid adaptation, training on small datasets, and reduced memory usage. In terms of model perplexity, prior work has shown diminished returns for class based language models constructed using very large training sets. This paper describes a method of using a class model as a backoff to a bigram model which produced … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 12 publications
0
2
0
Order By: Relevance
“…Related research, carried out independently and concurrently, has recently been reported in Miller and Alleva (1996). It describes a method that allows backoffs to occur from a word-to a category-based bigram language model.…”
Section: Backing-off From Word-to Category-based N-gramsmentioning
confidence: 99%
“…Related research, carried out independently and concurrently, has recently been reported in Miller and Alleva (1996). It describes a method that allows backoffs to occur from a word-to a category-based bigram language model.…”
Section: Backing-off From Word-to Category-based N-gramsmentioning
confidence: 99%
“…A better approach is to build a language model general enough to better estimate unseen and low frequency events, but specific enough to capture the ambiguous nature of words. A lot of work has been done on this, and the widely used techniques are interpolating the class-based LMs and word-based LMs [4,5,6], and backing-off from word-based LMs to class-based LMs when estimating probabilities of unseen events [7,8]. Performance of the LMs depends on the number of clusters to certain extent.…”
Section: Introductionmentioning
confidence: 99%