2020
DOI: 10.1101/2020.12.02.403477
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Thinking ahead: spontaneous prediction in context as a keystone of language in humans and machines

Abstract: Departing from rule-based linguistic models, advances in deep learning resulted in a new type of autoregressive deep language models (DLMs). These models are trained using a self-supervised next word prediction task. We provide empirical evidence for the connection between autoregressive DLMs and the human language faculty using spoken narrative and electrocorticographic recordings. Behaviorally, we demonstrate that humans have a remarkable capacity for word prediction in natural contexts, and that, given a su… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

8
61
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 40 publications
(69 citation statements)
references
References 101 publications
(232 reference statements)
8
61
0
Order By: Relevance
“…First, we found that the models’ performance on the next-word prediction task, but not other language tasks, is correlated with neural predictivity (see (Gauthier & Levy, 2019) for related evidence of fine-tuning of one model on tasks other than next-word-prediction leading to worse model-to-brain fit). Recent preprints conceptually replicate and extend this basic finding (Caucheteux & King, 2020; Goldstein et al, 2020; Wehbe et al, 2020; Wilcox et al, 2020). Language modeling (predicting the next word) is the task of choice in the natural language processing (NLP) community: it is simple, unsupervised, scalable, and appears to produce the most generally useful, successful language representations.…”
Section: Discussionmentioning
confidence: 69%
“…First, we found that the models’ performance on the next-word prediction task, but not other language tasks, is correlated with neural predictivity (see (Gauthier & Levy, 2019) for related evidence of fine-tuning of one model on tasks other than next-word-prediction leading to worse model-to-brain fit). Recent preprints conceptually replicate and extend this basic finding (Caucheteux & King, 2020; Goldstein et al, 2020; Wehbe et al, 2020; Wilcox et al, 2020). Language modeling (predicting the next word) is the task of choice in the natural language processing (NLP) community: it is simple, unsupervised, scalable, and appears to produce the most generally useful, successful language representations.…”
Section: Discussionmentioning
confidence: 69%
“…Crucially for us, the next-word probability estimates produced by these models was shown to be an impressive predictor of a host of indices of behavioral performance and brain activation, surpassing preceding generations of neural network models (reading times: Merkx & Frank, 2020;Wilcox et al, 2020;fMRI data: Schrimpf et al, 2020;MEG data: Caucheteux & King, 2020;N400 amplitudes: Merkx & Frank, 2020;Heilbron et al, 2021;ECOG activity: Goldstein et al, 2020;Schrimpf et al, 2020). Importantly, the improvement in linguistic accuracy (i.e., how good a model is at predicting the next word) converges with the model's "psychological" accuracy, i.e., how well it predicts context-based facilitations as measured in reading behavior and in the brain (Caucheteux & King, 2020;Goldstein et al, 2020;Schrimpf et al, 2020). Moreover, it is particularly the model's accuracy in the next-word prediction task (and not any other index of the model's performance) that explains its success in modeling brain activity (Schrimpf et al, 2020), which may indicate that-at least at the computational level-the brain pursues the same objective as the models trained in the next-word prediction task.…”
Section: Current Studymentioning
confidence: 99%
“…Moreover, the improvement of dynamical connectivity will also benefit the study of effective connectivity, causal relations between temporal signals, which should also greatly help clarify the "how" of language memory operations are performed by the brain and their failure. Finally, computational methods and artificial neural networks are also promising tools that, coupled with functional brain markers, can allow a better understanding of the computations and algorithms involved in natural language processing as well as their neural implementation (e.g., Goldstein et al, 2021;Jain & Huth, 2018, for two examples on word prediction in natural context; Martin, 2020, for a proposed architecture of the hierarchical and compositional structure of language, based on neurobiological and neurocomputational modeling evidence).…”
Section: Leveraging Advances In Functional Connectomicsmentioning
confidence: 99%