2020
DOI: 10.31234/osf.io/hs4ra
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Transformer Networks of Human Conceptual Knowledge

Abstract: We present a computational model capable of simulating aspects of human knowledge for thousands of real-world concepts. Our approach involves fine-tuning a transformer network for natural language understanding on participant-generated feature norms. We show that such a model can successfully extrapolate from its training dataset, and predict human knowledge for novel concepts and features. We also apply our model to stimuli from twenty-three previous experiments in semantic cognition research, and show that i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 70 publications
1
6
0
Order By: Relevance
“…As Figure 2C shows, the GPT-3 similarity ratings are correlated to some extent with human ratings. This is consistent with previous work suggesting that the internal representations of TLMs can be used to make reasonable predictions about human similarity judgments (Bhatia & Richie, 2021). GPT-3 accounts for some superordinate categories better than others, with correlations ranging between 0.16 (fish) and 0.58 (professions).…”
Section: Distinguishing Representation From Reasoningsupporting
confidence: 91%
See 1 more Smart Citation
“…As Figure 2C shows, the GPT-3 similarity ratings are correlated to some extent with human ratings. This is consistent with previous work suggesting that the internal representations of TLMs can be used to make reasonable predictions about human similarity judgments (Bhatia & Richie, 2021). GPT-3 accounts for some superordinate categories better than others, with correlations ranging between 0.16 (fish) and 0.58 (professions).…”
Section: Distinguishing Representation From Reasoningsupporting
confidence: 91%
“…For psychologists, property induction is relevant to a literature that assesses TLMs and predecessors such as LSA (Landauer & Dumais, 1997) as computational accounts of the acquisition, use, and representation of semantic knowledge. Recent work has evaluated the extent to which TLMs account for human similarity ratings, typicality ratings, and response times (Bhatia & Richie, 2021;Lake & Murphy, 2021), but there has been relatively little work on inductive reasoning. A notable exception is the work of Misra, Ettinger, and Taylor Rayz (2021), who focus on typicality and include property induction as one of the tasks that they consider.…”
Section: Introductionmentioning
confidence: 99%
“…As the first step toward automating analogical mapping, we adopt semantic representations of individual words generated by a machine-learning model, Word2vec (Mikolov et al, 2013). Word2vec and similar models based on distributional semantics, such as Global Vectors (GloVe; Pennington et al, 2014) and Bidirectional Encoder Representations from Transformers (BERT; Devlin et al, 2019), have proved successful in predicting behavioral judgments of lexical similarity or association (Hill et al, 2015; Hofmann et al, 2018; Pereira et al, 2016; Richie & Bhatia, 2021), neural responses to word and relation meanings (Huth et al, 2016; Pereira et al, 2018; Zhang et al, 2020), and high-level inferences including assessments of probability (Bhatia, 2017; Bhatia et al, 2019) and semantic verification (Bhatia & Richie, in press). In the simulations reported here, the semantic meanings of individual concepts are represented by 300-dimensional embeddings created by Word2vec after training on a corpus of articles drawn from Google News.…”
Section: Forming Representations Of Word Meanings and Semantic Relationsmentioning
confidence: 99%
“…Common optimization tasks for pretraining transformers, such as the masked LM task (Devlin et al, 2018) are quite similar to the word prediction tasks that are known to predict children's performance on other linguistic skills (Borovsky et al, 2012;Neuman et al, 2011;Gambi et al, 2020). Finally, TLMs tend to outperform other LMs in recent work modeling human reading times, eye-tracking data, and other psychological and psycholinguistic phenomena (Merkx and Frank, 2021;Schrimpf et al, 2020b,a;Hao et al, 2020;Bhatia and Richie, 2020;.…”
Section: Related Workmentioning
confidence: 98%