2021
DOI: 10.48550/arxiv.2109.07830
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Reframing Instructional Prompts to GPTk's Language

Abstract: How can model designers turn task instructions into effective prompts for language models? Backed by extensive empirical analysis on GPT3, we observe important features for successful instructional prompts, and propose several reframing techniques for model designers to create such prompts. For example, a complex task can be decomposed into multiple simpler tasks (Fig. 1a). We experiment over 12 NLP tasks across 6 diverse categories (question generation, classification, etc.). Our results show that reframing i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
34
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
1

Relationship

3
5

Authors

Journals

citations
Cited by 22 publications
(34 citation statements)
references
References 18 publications
0
34
0
Order By: Relevance
“…Prior work proposes better ways of formulating the problem (Zhao et al, 2021;Holtzman et al, 2021;Min et al, 2021a), better ways of choosing labeled examples for the demonstrations (Liu et al, 2021;Lu et al, 2021;Rubin et al, 2021), metatraining with an explicit in-context learning objective Min et al, 2021b), and learning to follow instructions as a variant of incontext learning (Mishra et al, 2021b;Efrat and Levy, 2020;Wei et al, 2022;Sanh et al, 2022). At the same time, some work reports brittleness and over-sensitivity for in-context learning (Lu et al, 2021;Zhao et al, 2021;Mishra et al, 2021a).…”
Section: Related Workmentioning
confidence: 99%
“…Prior work proposes better ways of formulating the problem (Zhao et al, 2021;Holtzman et al, 2021;Min et al, 2021a), better ways of choosing labeled examples for the demonstrations (Liu et al, 2021;Lu et al, 2021;Rubin et al, 2021), metatraining with an explicit in-context learning objective Min et al, 2021b), and learning to follow instructions as a variant of incontext learning (Mishra et al, 2021b;Efrat and Levy, 2020;Wei et al, 2022;Sanh et al, 2022). At the same time, some work reports brittleness and over-sensitivity for in-context learning (Lu et al, 2021;Zhao et al, 2021;Mishra et al, 2021a).…”
Section: Related Workmentioning
confidence: 99%
“…The release of GPT-3 (Brown et al, 2020) sparked a lot of excitement about the emergent ability of LMs in following discrete natural language prompts. Consequently, countless follow-up studies have used manually-designed discrete prompts for probing LMs (Petroni et al, 2019;Jiang et al, 2020), improving LMs few-shot ability (Schick and Schütze, 2021;Gao et al, 2021;Le Scao and Rush, 2021), and their zero-shot ability as well as transferability (Mishra et al, 2021a;Reynolds and McDonell, 2021). While discrete prompts have clear advantages, in addition to being human-readable and thus easily interpretable, we do not have efficient and algorithmic ways of reconstructing them.…”
Section: Related Workmentioning
confidence: 99%
“…For example, Shin et al (2020)'s algorithm discovers discrete prompts, alas the results are not human readable. Prior work also finds that model performance is highly sensitive to small changes in wordings (Mishra et al, 2021a), and optimization over the discrete space is non-trivial and often highly unstable. Our findings here about the disconnect between continuous prompts and their discrete interpretation provides another perspective on the difficulty of discovering discrete prompts via continuous optimizations algorithms that (directly or indirectly) leverage the continuous space (more discussion in §6).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…This hard limit in the number of tokens is problematic in tasks where examples include structured (e.g., MWoZ) or unstructured data (e.g., WoW, WiT, MSC) since each shot in the prompt requires thousands of tokens. To overcome this challenge, there are several possible alternatives to be explored: 1) improving the task description (Mishra et al, 2021;Reynolds and McDonell, 2021; rather than increasing the number of shots, 2) using prompt tuning (Li and Liang, 2021;Lester et al, 2021;Logan IV et al, 2021), where more shots would help since prompts are trained continuous embeddings, 3) using adapters tuning (Houlsby et al, 2019;, 4) converting the task into Table 13: Response to offensive language (YEA-SAYER (ELIZA) EFFECT test). The BST , DialoGPT , GPT-2 (Radford et al, 2019), and Kuki results are taken from .…”
Section: Limited Number Of Shotsmentioning
confidence: 99%