2023
DOI: 10.1101/2023.04.16.537080
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Driving and suppressing the human language network using large language models

Abstract: Transformer language models are today's most accurate models of language processing in the brain. Here, using fMRI-measured brain responses to 1,000 diverse sentences, we develop a GPT-based encoding model to identify new sentences that are predicted to drive or suppress responses in the human language network. We demonstrate that these model-selected 'out-of distribution' sentences indeed drive and suppress activity of human language areas in new individuals (85.7% increase and 97.5% decrease relative to the … Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
9
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 9 publications
(11 citation statements)
references
References 250 publications
2
9
0
Order By: Relevance
“…The overall predictivity was lower in the RH than the LH language fROIs (p <<0.0001 for all models in Experiment 1 and all checkpoints in Experiment 2), in line with past findings (e.g. (Schrimpf et al, 2021; Tuckute et al, 2023)).…”
Section: Resultssupporting
confidence: 90%
See 2 more Smart Citations
“…The overall predictivity was lower in the RH than the LH language fROIs (p <<0.0001 for all models in Experiment 1 and all checkpoints in Experiment 2), in line with past findings (e.g. (Schrimpf et al, 2021; Tuckute et al, 2023)).…”
Section: Resultssupporting
confidence: 90%
“…The fROIs are identified with an independent localizer, as described in Methods . The general pattern of results (presented in main Figure 2) holds across hemispheres (although predictivity is higher in the LH, in line with other work; e.g., Schrimpf et al, 2021; Tuckute et al, 2023) and fROIs.…”
Section: Supplementary Figuresupporting
confidence: 88%
See 1 more Smart Citation
“…It will also be important to use additional means of model evaluation, such as model-matched stimuli 25, 27, 57 , stimuli optimized for a model’s predicted response 110, 130,131,132 , directly substituting brain responses into models 112 , or recently proposed alternative methods to measure representational similarity 111 . These additional types of evaluations could help address some of the limitations discussed in the previous section.…”
Section: Discussionmentioning
confidence: 99%
“…While the naturalistic nature of these stimuli means that we did not necessarily have repeated presentation of the same word(s) across stories, we can use natural language processing (NLP) techniques to group words into clusters of semantically related words and use the clusters to help understand why concrete representations are more reliable, even when generalizing over individual words and concepts. Numerous recent studies have demonstrated parallels in language representation between NLP models and human neural processing 13,[64][65][66][67] . Here, we used a word-embedding NLP model (GloVe) 52 to understand how the semantic relationships among concrete and abstract words relate to the reliability of their neural representations.…”
Section: Stable Clusters Of Concrete Words Drive Reliability Of Repre...mentioning
confidence: 99%