2018
DOI: 10.31234/osf.io/g9j83
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Semantic representations extracted from large language corpora predict high-level human judgment in seven diverse behavioral domains

Abstract: Recent advances in machine learning, combined with the increased availability of large natural language datasets, have made it possible to uncover semantic representations that characterize what people know about and associate with a wide range of objects and concepts. In this paper, we examine the power of word embeddings, a popular approach for uncovering semantic representations, for studying high-level human judgment. Word embeddings are typically applied to linguistic and semantic tasks, however we show t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(12 citation statements)
references
References 21 publications
(31 reference statements)
1
11
0
Order By: Relevance
“…For example, among the 4436 words in the feature norms of Buchanan et al (2019), participants listed on average 15 features per word. Similarly, Richie, Zou, and Bhatia (2018) showed that predictive accuracy of models regressing numerical semantic judgments about words (e.g., tastiness of foods) onto high-dimensional word embedding-based vector representations of words only started to plateau with about ten principal components (within, e.g., the embeddings for the set of foods), suggesting that people were making their semantic judgments on the basis of many (>>>2) dimensions of the Fig. 1.…”
Section: The Spatial Arrangement Methods For Measuring Similaritymentioning
confidence: 92%
“…For example, among the 4436 words in the feature norms of Buchanan et al (2019), participants listed on average 15 features per word. Similarly, Richie, Zou, and Bhatia (2018) showed that predictive accuracy of models regressing numerical semantic judgments about words (e.g., tastiness of foods) onto high-dimensional word embedding-based vector representations of words only started to plateau with about ten principal components (within, e.g., the embeddings for the set of foods), suggesting that people were making their semantic judgments on the basis of many (>>>2) dimensions of the Fig. 1.…”
Section: The Spatial Arrangement Methods For Measuring Similaritymentioning
confidence: 92%
“…The CSLB norms used in Sommerauer and Fokkens's (2018) study contained only perceptual, taxonomic, and other properties mainly for concrete concepts. By contrast, most of the properties analyzed in Grand et al (2018) and Richie et al (2019) were abstract. Therefore, in these studies, concrete and abstract properties were not compared together, and thus it still remains unclear which properties can be captured better than others by DSMs.…”
Section: Related Workmentioning
confidence: 94%
“…Using GloVe vectors, they revealed that abstract properties, such as gender and danger, were significantly predicted across all categories, and even perceptual properties, such as size, were captured for some relevant categories. Richie et al (2019) also showed that 14 properties, most of which are abstract (e.g., competent and sincere), were predicted by skip‐gram vectors.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…(vehicles not in the testing set). By contrast, prior work using projection techniques to predict feature ratings from embedding spaces (Grand et al, 2018;Richie et al, 2019) has used adjectives as endpoints, ignoring the potential influence of domain-level semantic context on similarity judgments (e.g., "size" was defined as a vector from "small," "tiny," "minuscule" to "large," "huge," "giant," regardless of semantic context). However, as we argued above, feature ratings may be impacted by semantic context much as-and perhaps for the same reasons as-similarity judgments.…”
Section: Experiments 2: Contextual Projection Captures Reliable Infor...mentioning
confidence: 99%