Proceedings of the Workshop on Human Language Technology - HLT '93 1993
DOI: 10.3115/1075671.1075730
|View full text |Cite
|
Sign up to set email alerts
|

Corpus-based statistical sense resolution

Abstract: The three corpus-based statistical sense resolution methods studied here attempt to infer the correct sense of a polysemous word by using knowledge about patterns of word cooccurrences. The techniques were based on Bayesian decision theory, neural networks, and content vectors as used in information retrieval. To understand these methods better, we posed s very specific problem: given a set of contexts, each containing the noun line in a known sense, construct a classifier that selects the correct sense of lin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
55
4
2

Year Published

1997
1997
2018
2018

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 82 publications
(67 citation statements)
references
References 8 publications
3
55
4
2
Order By: Relevance
“…Obviously, large values of window sizes capture dependencies at longer range but also dilute the effect of the words closer to the term. Leacock et al [28] used a window size of 50, while Yarowsky [29] argued that a small window size of 3 or 4 had better performance. A small window size has an advantage of requiring less system space and running time [30].…”
Section: Data Acquisition and Feature Extractionmentioning
confidence: 99%
“…Obviously, large values of window sizes capture dependencies at longer range but also dilute the effect of the words closer to the term. Leacock et al [28] used a window size of 50, while Yarowsky [29] argued that a small window size of 3 or 4 had better performance. A small window size has an advantage of requiring less system space and running time [30].…”
Section: Data Acquisition and Feature Extractionmentioning
confidence: 99%
“…While n-grams perform well in partof-speech tagging and speech processing, they require a fixed interdependency structure that is inappropriate for the broad class of contextual features used in word-sense disambiguation. However, the Naive Bayes classifier has been found to perform well for word-sense disambiguation both here and in a variety of other works (e.g., (Bruce and Wiebe, 1994a), (Gale et al, 1992), (Leacock et al, 1993), and (Mooney, 1996)). …”
Section: Related Workmentioning
confidence: 99%
“…In particular, we use subsets of the line data (Leacock et al, 1993) and the English lexical sample data from the SENSEVAL-2 comparative exercise among word sense disambiguation systems (Edmonds and Cotton, 2001). …”
Section: Experimental Methodologymentioning
confidence: 99%