Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing 2015
DOI: 10.18653/v1/d15-1002
|View full text |Cite
|
Sign up to set email alerts
|

Distributional vectors encode referential attributes

Abstract: Distributional methods have proven to excel at capturing fuzzy, graded aspects of meaning (Italy is more similar to Spain than to Germany). In contrast, it is difficult to extract the values of more specific attributes of word referents from distributional representations, attributes of the kind typically found in structured knowledge bases (Italy has 60 million inhabitants). In this paper, we pursue the hypothesis that distributional vectors also implicitly encode referential attributes. We show that a standa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
59
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 55 publications
(62 citation statements)
references
References 21 publications
(16 reference statements)
0
59
0
Order By: Relevance
“…Early work in probing, (also known as diagnostic classification ,) extracted properties like parts-of-speech, gender, tense, and number from distributional word vector spaces like word2vec and GloVe (Mikolov et al, 2013;Pennington et al, 2014) using linear classifiers (Köhn, Part-of- 2015; Gupta et al, 2015). Soon after, the investigation of intermediate layers of deep models using linear probes was introduced independently by Ettinger et al (2016) and Shi et al (2016) in NLP and Alain and Bengio (2016) in computer vision.…”
Section: Related Workmentioning
confidence: 99%
“…Early work in probing, (also known as diagnostic classification ,) extracted properties like parts-of-speech, gender, tense, and number from distributional word vector spaces like word2vec and GloVe (Mikolov et al, 2013;Pennington et al, 2014) using linear classifiers (Köhn, Part-of- 2015; Gupta et al, 2015). Soon after, the investigation of intermediate layers of deep models using linear probes was introduced independently by Ettinger et al (2016) and Shi et al (2016) in NLP and Alain and Bengio (2016) in computer vision.…”
Section: Related Workmentioning
confidence: 99%
“…There is some empirical evidence that distributional data can be used for inferring properties in Johns & Jones 2012, Fȃgȃrȃşan, Vecchi & Clark 2015, Gupta et al 2015, and Herbelot & Vecchi 2015. They test whether distributional vectors can be used to predict a word's properties (where, as above, I use the term "properties of a word" to mean properties that apply to all entities in the word's extension).…”
Section: :20mentioning
confidence: 99%
“…Finally, there is a need for more research on distributional models for property inference, to develop efficient models beyond the initial approaches proposed by Johns & Jones (2012), Fȃgȃrȃşan, Vecchi & Clark (2015), Herbelot & Vecchi (2015) and Gupta et al (2015) and to see what kinds of properties can be reliably learned and whether verb properties can be learned as well as noun properties.…”
Section: Katrin Erkmentioning
confidence: 99%
“…But it implements the mapping as a systematic linear transformation. Our approach is similar to Gupta et al (2015), who predict numerical attributes for unseen concepts (countries and cities) from distributional vectors, getting comparably accurate estimates for features such as the GDP or CO 2 emissions of a country. We complement such research by providing a more formal interpretation of the mapping between language and world knowledge.…”
Section: Generalised Quantifiersmentioning
confidence: 99%