1997
DOI: 10.21236/ada640606
|View full text |Cite
|
Sign up to set email alerts
|

Structure and Performance of a Dependency Language Model

Abstract: We present a maximum entropy language model that incorporates both syntax and semantics via a dependency grammar. Such a grammar expresses the relations between words by a directed graph. Because the edges of this graph may connect words that are arbitrarily far apart in a sentence, this technique can incorporate the predictive p o wer of words that lie outside of bigram or trigram range. We h a ve built several simple dependency models, as we call them, and tested them in a speech recognition experiment. We r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
53
0

Year Published

1999
1999
2016
2016

Publication Types

Select...
4
3
1

Relationship

3
5

Authors

Journals

citations
Cited by 50 publications
(53 citation statements)
references
References 6 publications
0
53
0
Order By: Relevance
“…How these constraints are incorporated varies from estimating -gram probabilities from grammar-generated data [70] to computing a linear interpolation of the two models [43]. Most recently, syntactic information has been used specifically to determine equivalence classes on the -gram history, resulting in so-called dependency language models [19], [56], sometimes also referred to as structured language models [20], [42], [66].…”
Section: B Syntactically Driven Span Extensionmentioning
confidence: 99%
“…How these constraints are incorporated varies from estimating -gram probabilities from grammar-generated data [70] to computing a linear interpolation of the two models [43]. Most recently, syntactic information has been used specifically to determine equivalence classes on the -gram history, resulting in so-called dependency language models [19], [56], sometimes also referred to as structured language models [20], [42], [66].…”
Section: B Syntactically Driven Span Extensionmentioning
confidence: 99%
“…As shown in Table 4, although the size of CM (i.e., # of dependencies) is much larger, the improvement is very limited. On the other end of the spectrum, we have models that use sophisticated syntactic structure, such as dependencybased models [4,5] and constituency-based models [2,3]. They all use syntactic grammars for parsing and the parsing model is estimated from a manually annotated training data (i.e.…”
Section: Discussion On Term Dependencies With or Without Linguistic mentioning
confidence: 99%
“…[4]), we assume that the sum ∑ L P(Q, L|D) over all the possible Ls is dominated by a single term L * which is the most probable linkage of the query Q. Below we simply use L to represent L * .…”
Section: A New Modelmentioning
confidence: 99%
“…Jurafsky et al [71] use stochastic CFGs (SCFGs) to extend the corpus for training and interpolates SCFG probabilities with bi-gram probabilities. Chelba et al [24] use a dependency grammar framework with maximum entropy models to constrain word prediction by the linguistically related words in the past. The most important instance of LMs that use syntactic structure is presented in [25].…”
Section: Language Modelmentioning
confidence: 99%
“…The first attempts are based on using context free grammars (CFGs) [27,137,71]. The main contribution of structured LMs is started with Chelba et al [24] in which a dependency grammar framework with maximum entropy models is used to constrain the word prediction by the linguistically related words in the past.…”
Section: Structured Lmsmentioning
confidence: 99%