Proceedings of the Eighth Conference on European Chapter of the Association for Computational Linguistics - 1997
DOI: 10.3115/979617.979665
|View full text |Cite
|
Sign up to set email alerts
|

A model of lexical attraction and repulsion

Abstract: This paper introduces new methods based on exponential families for modeling the correlations between words in text and speech. While previous work assumed the effects of word co-occurrence statistics to be constant over a window of several hundred words, we show that their influence is nonstationary on a much smaller time scale. Empirical data drawn from English and Japanese text, as well as conversational speech, reveals that the "attraction" between words decays exponentially, while stylistic and syntactic … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

1999
1999
2014
2014

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 26 publications
(28 citation statements)
references
References 3 publications
0
28
0
Order By: Relevance
“…At the simplest level, if the algorithm is allowed to use a longer text segment around a seed word, a larger set of terms is likely to be measured. More interesting, in language usage, at least for English, the tendency is to avoid repeating a word in an adjacent sentence and to use a replacement term, such as a synonym (see, e.g., Beeferman et al, 1997). The relevancy scores of the words that are also seen in the one-sentence lists are, for the most part, higher in the three-sentence lists.…”
Section: Stabilitymentioning
confidence: 99%
See 1 more Smart Citation
“…At the simplest level, if the algorithm is allowed to use a longer text segment around a seed word, a larger set of terms is likely to be measured. More interesting, in language usage, at least for English, the tendency is to avoid repeating a word in an adjacent sentence and to use a replacement term, such as a synonym (see, e.g., Beeferman et al, 1997). The relevancy scores of the words that are also seen in the one-sentence lists are, for the most part, higher in the three-sentence lists.…”
Section: Stabilitymentioning
confidence: 99%
“…In essence, a word can be defined by its context in usage. Beeferman and colleagues observed that words tend to correlate with other words over a certain range within the text stream (Beeferman, Berger, & Lafferty, 1997). Computational linguists have also exploited this aspect of language-for word sense disambiguation, as a particular example (Yarowsky, 1995).…”
Section: Foundations Of the Techniquementioning
confidence: 99%
“…Not only can the Gaussian prior be applied to maximum entropy modeling, but it can also be applied in the more general minimum divergence paradigm [37,38]. Maximizing entropy is equivalent to finding the model with the smallest Kullback-Leibler divergence from the uniform distribution.…”
Section: Discussionmentioning
confidence: 99%
“…The method we use here, described in Beeferman, Berger, and Lafferty (1997), starts with a trigram model as a prior, or default distribution, and tacks onto the model a set of features to account for the long-range lexical properties of language. The features are trigger pairs, automatically discovered by analyzing a corpus of text using a mutual information heuristic Figure 2.…”
Section: Some Doctors Are More Skilled At Doing the Procedures Than Otmentioning
confidence: 99%