2021
DOI: 10.48550/arxiv.2110.07298
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

LFPT5: A Unified Framework for Lifelong Few-shot Language Learning Based on Prompt Tuning of T5

Abstract: Existing approaches to lifelong language learning rely on plenty of labeled data for learning a new task, which is hard to obtain in most real scenarios. Considering that humans can continually learn new tasks from a handful of examples, we expect the models also to be able to generalize well on new few-shot tasks without forgetting the previous ones. In this work, we define this more challenging yet practical problem as Lifelong Few-shot Language Learning (LFLL) and propose a unified framework for it based on… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(16 citation statements)
references
References 46 publications
(43 reference statements)
0
16
0
Order By: Relevance
“…We run extensive experiments on standard CL benchmarks for text classification, and show that Progressive Prompts outperforms state-of-the-art approaches on both BERT and T5 architectures (Devlin et al, 2018;Raffel et al, 2020). We show over 20% improvement over the current SOTA for T5 model (Qin & Joty, 2021). Furthermore, we run experiments on a more challenging CL setup with longer task sequences, and show that our method outperforms prior approaches for T5 and BERT architectures.…”
Section: Introductionmentioning
confidence: 79%
See 3 more Smart Citations
“…We run extensive experiments on standard CL benchmarks for text classification, and show that Progressive Prompts outperforms state-of-the-art approaches on both BERT and T5 architectures (Devlin et al, 2018;Raffel et al, 2020). We show over 20% improvement over the current SOTA for T5 model (Qin & Joty, 2021). Furthermore, we run experiments on a more challenging CL setup with longer task sequences, and show that our method outperforms prior approaches for T5 and BERT architectures.…”
Section: Introductionmentioning
confidence: 79%
“…The training objective for task T k (k ∈ {1...m}) is to find prompt parameters θ P k that minimize the negative log probability of training examples under our progressive prompt and the frozen base model: Large number of tasks While previous CL approaches mostly focused on relatively short task sequences of 3-5 tasks (Huang et al, 2021;de Masson D'Autume et al, 2019;Qin & Joty, 2021), a more realistic CL scenario would be long sequences encompassing more tasks. Hence, we create a benchmark of 15 text classification tasks to evaluate the performance of Progressive Prompts along with widely adopted CL approaches.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Along with continual learning, the parameters of big models are also continually updated, which means that those prompts tuned based on historic parameters may fail based on the latest version of big models. Therefore, it is meaningful to ensure that the prompts tuned on historic parameters can continue to work on new models, which has been demonstrated in the paper [572].…”
Section: Continual Learningmentioning
confidence: 98%