2005
DOI: 10.1162/089120105775299122
|View full text |Cite
|
Sign up to set email alerts
|

Co-occurrence Retrieval: A Flexible Framework for Lexical Distributional Similarity

Abstract: Techniques that exploit knowledge of distributional similarity between words have been proposed in many areas of Natural Language Processing. For example, in language modeling, the sparse data problem can be alleviated by estimating the probabilities of unseen co-occurrences of events from the probabilities of seen co-occurrences of similar events. In other applications, distributional similarity is taken to be an approximation to semantic similarity. However, due to the wide range of potential applications an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
100
0
1

Year Published

2008
2008
2022
2022

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 116 publications
(102 citation statements)
references
References 29 publications
(70 reference statements)
1
100
0
1
Order By: Relevance
“…Automated approaches for representing the semantic content of terms and similarity and relatedness between them have been widely used in a number of Natural Language Processing (NLP) applications in both general English (Budanitsky and Hirst, 2006;Landauer, 2006;Resnik, 1999;Weeds and Weir, 2005) and specialized terminological domains such as bioinformatics (Ferreira et al, 2013;Lord et al, 2003;Mazandu et al, 2016;Wang et al, 2007;Yang et al, 2012) and medicine (Garla and Brandt, 2012;Lee et al, 2008;Liu et al, 2012;Pakhomov et al, 2010;Pedersen et al, 2007;Sajadi, 2014). A subset of these methods, distributional semantics, relies on the co-occurrence information between words obtained from large corpora of text and makes the assumption that words with similar or related meanings tend to occur in similar contexts.…”
Section: Introductionmentioning
confidence: 99%
“…Automated approaches for representing the semantic content of terms and similarity and relatedness between them have been widely used in a number of Natural Language Processing (NLP) applications in both general English (Budanitsky and Hirst, 2006;Landauer, 2006;Resnik, 1999;Weeds and Weir, 2005) and specialized terminological domains such as bioinformatics (Ferreira et al, 2013;Lord et al, 2003;Mazandu et al, 2016;Wang et al, 2007;Yang et al, 2012) and medicine (Garla and Brandt, 2012;Lee et al, 2008;Liu et al, 2012;Pakhomov et al, 2010;Pedersen et al, 2007;Sajadi, 2014). A subset of these methods, distributional semantics, relies on the co-occurrence information between words obtained from large corpora of text and makes the assumption that words with similar or related meanings tend to occur in similar contexts.…”
Section: Introductionmentioning
confidence: 99%
“…Entailment and inference systems can improve sentence-level entailment resolution by detecting the presence and direction of wordlevel hyponymy relations. Distributionally similar words have been used for smoothing language models and word co-occurrence probabilities (Dagan et al, 1999;Weeds and Weir, 2005), and hyponyms can be more suitable for this application.…”
Section: Introductionmentioning
confidence: 99%
“…Relying on such lexical acquisition methods, automatically induced features and relations can complement directly accessible ones. In a second step, it will often be necessary to bridge non-identical but related properties; this task can be approached using standard approaches to semantic similarity, such as distributional measures (Dagan et al, 1999;Curran, 2004;Weeds and Weir, 2005;Budanitsky and Hirst, 2006;Padó and Lapata, 2007). Once a set of (comparable) rich features is available, we assume that an automation of the verb class linking is relatively straightforward.…”
Section: Resultsmentioning
confidence: 99%