Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing 2022
DOI: 10.18653/v1/2022.emnlp-main.340
|View full text |Cite
|
Sign up to set email alerts
|

Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(13 citation statements)
references
References 0 publications
0
11
0
Order By: Relevance
“…CrossFit (Ye et al, 2021). To investigate models' few-shot learning capabilities across tasks, a collection of 269 NLP task datasets, known as CrossFit, has been assembled, covering 13 task types (Wang et al, 2022). In addition to being used for instruction fine-tuning, this dataset is employed for studying models' cross-task generalization and transfer learning abilities.…”
Section: Collection and Improvement Of Existing Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…CrossFit (Ye et al, 2021). To investigate models' few-shot learning capabilities across tasks, a collection of 269 NLP task datasets, known as CrossFit, has been assembled, covering 13 task types (Wang et al, 2022). In addition to being used for instruction fine-tuning, this dataset is employed for studying models' cross-task generalization and transfer learning abilities.…”
Section: Collection and Improvement Of Existing Datasetsmentioning
confidence: 99%
“…Flan 2022 (Longpre et al, 2023a). The Flan 2022 dataset consists of five parts, namely Flan 2021, T0 (Victor et al, 2022), SUPER-NATURAL INSTRUCTIONS (Wang et al, 2022), CoT datasets, and Dialog datasets. It encompasses as many as 1836 tasks.…”
Section: Collection and Improvement Of Existing Datasetsmentioning
confidence: 99%
“…These efforts include benchmarks like BLUE [63], HunFlair [90], BLURB [20], and BigBio [16], which provide datasets and tasks for evaluating biomedical language understanding and reasoning. Additionally, there are biomedical datasets geared towards prompt-based learning and evaluation of few and zero-shot classification, such as Super-NaturalInstructions [89] and BoX [61]. Out of all benchmarks mentioned above, only BoX contains one CS dataset covering five SLRs, however, this dataset is private.…”
Section: Dataset Overlapmentioning
confidence: 99%
“…As generalisation to new domains (with limited in-domain annotation effort) is one of the main desiderata of TOD, some recent work on dialog NLU (Fuisz et al, 2022; has recognised that ID and VE can be cast as question answering (QA) tasks: this facilitates transfer from models trained on large QA datasets (Rajpurkar et al, 2016a;, allowing also to capitalise on other large datasets previously recast as QA (McCann et al, 2018;Wang et al, 2022b). These efforts, however, amount to sequential transfer with standard fine-tuning for QA and thus (i) do not align their fine-tuning with the models' pretraining objective; and without an LM-based objective they (ii) cannot benefit from cross-task transfer via natural language task formulations.…”
Section: Introductionmentioning
confidence: 99%