Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.407
|View full text |Cite
|
Sign up to set email alerts
|

Improving and Simplifying Pattern Exploiting Training

Abstract: Recently, pre-trained language models (LMs) have achieved strong performance when finetuned on difficult benchmarks like Super-GLUE. However, performance can suffer when there are very few labeled examples available for fine-tuning. Pattern Exploiting Training (PET) is a recent approach that leverages patterns for few-shot learning. However, PET uses task-specific unlabeled data. In this paper, we focus on few shot learning without any unlabeled data and introduce ADAPET, which modifies PET's objective to prov… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(36 citation statements)
references
References 11 publications
0
36
0
Order By: Relevance
“…These prompts share the same format of masked language modeling, the pre-training tasks of many pre-trained LMs, and thus leads to improved few-shot performance. Extending from PET, Gao et al (2020) proposed LM-BFF which learns to generate prompts automatically and incorporates demonstrations into the input; Tam et al (2021) proposed ADAPET which densifies the supervision signal with a label conditioning objective.…”
Section: Related Workmentioning
confidence: 99%
“…These prompts share the same format of masked language modeling, the pre-training tasks of many pre-trained LMs, and thus leads to improved few-shot performance. Extending from PET, Gao et al (2020) proposed LM-BFF which learns to generate prompts automatically and incorporates demonstrations into the input; Tam et al (2021) proposed ADAPET which densifies the supervision signal with a label conditioning objective.…”
Section: Related Workmentioning
confidence: 99%
“…A new fine-tuning methodology named prompt-tuning has arisen: adapting the pre-trained language model directly as a predictor through completion of a cloze task. Prompt-tuning for pre-trained language models is a rapidly emerging field in natural language processing [40,46,71] and have attracted lots of attention. Originally from GPT-3, prompt-tuning has been applied to various of tasks including relation extraction [20], event extraction [21,59], named entity recognition [5,7], entity typing [13], and so on.…”
Section: Prompt-tuningmentioning
confidence: 99%
“…While PET requires task-specific prompts, it achieves better performance than GPT-3 in-context with smaller models [26]. ADAPET improves upon PET by providing more supervision during fine-tuning [27]. LM-BFF [11] improves prompt-based fine-tuning by dynamically constructing prompts.…”
Section: Few-shot Learning In Nlpmentioning
confidence: 99%