2021
DOI: 10.48550/arxiv.2104.08786
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity

Abstract: When primed with only a handful of training samples, very large pretrained language models such as GPT-3, have shown competitive results when compared to fully-supervised fine-tuned large pretrained language models. We demonstrate that the order in which the samples are provided can be the difference between near state-of-the-art and random guess performance: Essentially some permutations are "fantastic" and some not. We analyse this phenomenon in detail, establishing that: it is present across model sizes (ev… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

2
56
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 51 publications
(72 citation statements)
references
References 11 publications
2
56
0
Order By: Relevance
“…In-context learning has been the focus of significant study since its introduction. Prior work proposes better ways of formulating the problem (Zhao et al, 2021;Holtzman et al, 2021;Min et al, 2021a), better ways of choosing labeled examples for the demonstrations (Liu et al, 2021;Lu et al, 2021;Rubin et al, 2021), metatraining with an explicit in-context learning objective Min et al, 2021b), and learning to follow instructions as a variant of incontext learning (Mishra et al, 2021b;Efrat and Levy, 2020;Wei et al, 2022;Sanh et al, 2022). At the same time, some work reports brittleness and over-sensitivity for in-context learning (Lu et al, 2021;Zhao et al, 2021;Mishra et al, 2021a).…”
Section: Related Workmentioning
confidence: 99%
“…In-context learning has been the focus of significant study since its introduction. Prior work proposes better ways of formulating the problem (Zhao et al, 2021;Holtzman et al, 2021;Min et al, 2021a), better ways of choosing labeled examples for the demonstrations (Liu et al, 2021;Lu et al, 2021;Rubin et al, 2021), metatraining with an explicit in-context learning objective Min et al, 2021b), and learning to follow instructions as a variant of incontext learning (Mishra et al, 2021b;Efrat and Levy, 2020;Wei et al, 2022;Sanh et al, 2022). At the same time, some work reports brittleness and over-sensitivity for in-context learning (Lu et al, 2021;Zhao et al, 2021;Mishra et al, 2021a).…”
Section: Related Workmentioning
confidence: 99%
“…It is well known that prompt-based methods are sensitive to the many aspects of prompts including contexts (Jiang et al, 2020;Shin et al, 2020), orders (Lu et al, 2021) and length , and inappropriate prompts will cause bad performance.…”
Section: Effects Of Prompt Lengthmentioning
confidence: 99%
“…Zhao et al (2021) propose to remove the model bias before using GPT-3, which not only increases the accuracy but also reduces the variance. Lu et al (2021) work on how to order the few labeled data as input of GPT-3 by constructing an artificial development set. One concurrent with our work, Yoo et al ( 2021) consider distilling knowledge from GPT-3 with synthetic data.…”
Section: Related Workmentioning
confidence: 99%