Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.244
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Task Generalization via Natural Language Crowdsourcing Instructions

Abstract: Humans (e.g., crowdworkers) have a remarkable ability in solving different tasks, by simply reading textual instructions that define them and looking at a few examples. Despite the success of the conventional supervised learning on individual datasets, such models often struggle with generalization across tasks (e.g., a question-answering system cannot solve classification tasks). A long-standing challenge in AI is to build a model that learns a new task by understanding the humanreadable instructions that def… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
70
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 86 publications
(109 citation statements)
references
References 33 publications
1
70
0
Order By: Relevance
“…Fine-tuned LMs for Instruction Learning Recent work shows that fine-tuning LMs to learn task instructions on a wide variety of tasks can further leverage the inductive bias of LMs to perform instruction learning (Zhong et al, 2021;Mishra et al, 2021;Wei et al, 2021). Our work is partially inspired by this line of work, but we work under the more generic few-shot meta-learning setting, and show that our approach out-performs both instruction tuning and existing few-shot meta-learning methods (e.g., MAML).…”
Section: Related Workmentioning
confidence: 99%
“…Fine-tuned LMs for Instruction Learning Recent work shows that fine-tuning LMs to learn task instructions on a wide variety of tasks can further leverage the inductive bias of LMs to perform instruction learning (Zhong et al, 2021;Mishra et al, 2021;Wei et al, 2021). Our work is partially inspired by this line of work, but we work under the more generic few-shot meta-learning setting, and show that our approach out-performs both instruction tuning and existing few-shot meta-learning methods (e.g., MAML).…”
Section: Related Workmentioning
confidence: 99%
“…The Instruction Paradigm Efrat and Levy [2020] propose to learn new tasks from natural language instructions. Mishra et al [2022] and collect crowdsourcing instructions used to create NLP datasets into a benchmark for measuring the ability to solve tasks by reading instructions.…”
Section: Related Workmentioning
confidence: 99%
“…MTL has also contributed to sub-fields of text generation, such as open-ended dialogue system [160,9], task-oriented dialogue system [131], text style transfer [16], and question answering [63]. At the same time, researchers explore the transferability of models trained on multi-task datasets [99]. FLAN [145], T0 [125], and ZeroPrompt [153] investigate the zero-shot generalization abilities of large PLMs trained on numerous datasets with well-designed prompts.…”
Section: Related Workmentioning
confidence: 99%
“…GPT-3 [14] further combines several demonstrations to input to learn task patterns, which is called in-context learning. Some researchers also design elaborate prompts or demonstrations for each task and dataset and investigate their effectiveness and robustness [145,125,153,99]. Even so, whether models can truly understand the semantic meanings of prompts looks worthy of further investigation [144].…”
Section: Related Workmentioning
confidence: 99%