2022 Conference on Cognitive Computational Neuroscience 2022
DOI: 10.32470/ccn.2022.1255-0
|View full text |Cite
|
Sign up to set email alerts
|

Reconstructing the cascade of language processing in the brain using the internal computations of transformer language models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 14 publications
(13 citation statements)
references
References 0 publications
1
10
0
Order By: Relevance
“…Probing the nature of these fits further, they find the best model-to-brain match obtains with the middle layers of the LLMs (e.g., layers 8-9 of a 12-layer feed-forward transformer network) rather than input or output layers. Kumar et al (2022) report similar results for fMRI data collected while participants listen to narratives.…”
Section: Predictability Processing and Llmssupporting
confidence: 67%
See 3 more Smart Citations
“…Probing the nature of these fits further, they find the best model-to-brain match obtains with the middle layers of the LLMs (e.g., layers 8-9 of a 12-layer feed-forward transformer network) rather than input or output layers. Kumar et al (2022) report similar results for fMRI data collected while participants listen to narratives.…”
Section: Predictability Processing and Llmssupporting
confidence: 67%
“…Kumar et al. (2022) report similar results for fMRI data collected while participants listen to narratives.…”
Section: Introductionmentioning
confidence: 59%
See 2 more Smart Citations
“…In such a mapping, early language areas will be better modeled by embedding extracted from early layers of DLMs, whereas higher-order areas will be better modeled by embeddings extracted from later layers of DLMs. Interestingly, studies that examined the layer-by-layer match between DLM embeddings and brain activity using fMRI have observed that intermediate layers tend to provide the best fit across many language ROIs (3, 15, 30, 31). These findings do not support the hypothesis that DLMs capture the processing sequence of words in natural language in the human brain.…”
Section: Introductionmentioning
confidence: 99%