2022
DOI: 10.1101/2022.02.22.481293
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generalization in Sensorimotor Networks Configured with Natural Language Instructions

Abstract: We present neural models of one of humans' most astonishing cognitive feats: the ability to interpret linguistic instructions in order to perform novel tasks with just a few practice trials. Models are trained on a set of commonly studied psychophysical tasks, and receive linguistic instructions embedded by transformer architectures pre-trained on natural language processing. Our best performing models can perform an unknown task with a performance of 80% correct on average based solely on linguistic instructi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 75 publications
(139 reference statements)
0
6
0
Order By: Relevance
“…The study of cognitive tasks through simulations in artificial networks has led to substantial advances in understanding neural computation in the past decade [9][10][11][12][13][14][15][16][17][18] . However, researchers typically trained artificial neural networks to perform single tasks in isolation, with few exceptions [19][20][21][22] , somewhat limiting the insights into biological neural circuits that perform many tasks. One exception to this trend is Yang et al 2019 20 , in which the authors trained a single network to perform twenty related tasks and thereby identified clustered representations in state space that supported task compositionality.…”
Section: Introductionmentioning
confidence: 99%
“…The study of cognitive tasks through simulations in artificial networks has led to substantial advances in understanding neural computation in the past decade [9][10][11][12][13][14][15][16][17][18] . However, researchers typically trained artificial neural networks to perform single tasks in isolation, with few exceptions [19][20][21][22] , somewhat limiting the insights into biological neural circuits that perform many tasks. One exception to this trend is Yang et al 2019 20 , in which the authors trained a single network to perform twenty related tasks and thereby identified clustered representations in state space that supported task compositionality.…”
Section: Introductionmentioning
confidence: 99%
“…Our work instantiates these concepts by learning by using gradient descent for latent updates. Similar use of gradient descent to update latent variables was proposed as model for schizophrenia by Yamashita and Tani (2012) and more recently to retrieve language instructions for cognitive tasks (Riveland and Pouget, 2022). Our work uses these methods as a component in a continually learning agent to learn disentangled representations.…”
Section: Related Workmentioning
confidence: 97%
“…Several recent studies in neuroscience have applied analytic tools to identify the neural basis of rapid generalization in biological neural networks. Such studies employed various measures -crosscondition generalization [2,40,8,4], state-space projections of task-related compositional codes [51,43,25], and Parallelism Score [2] -to quantify the generalizability and abstraction of representations. Prior work in neuroscience has primarily evaluated compositionality in limited context settings (e.g., up to 10 contexts), or without manipulating different types of features (e.g., higher-order vs. sensory/motor features).…”
Section: Related Workmentioning
confidence: 99%