Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.469
|View full text |Cite
|
Sign up to set email alerts
|

Diverse Distributions of Self-Supervised Tasks for Meta-Learning in NLP

Abstract: Meta-learning considers the problem of learning an efficient learning process that can leverage its past experience to accurately solve new tasks. However, the efficacy of meta-learning crucially depends on the distribution of tasks available for training, and this is often assumed to be known a priori or constructed from limited supervised datasets. In this work, we aim to provide task distributions for meta-learning by considering self-supervised tasks automatically proposed from unlabeled text, to enable la… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 37 publications
0
2
0
Order By: Relevance
“…The objective of assuming this distribution is to sample T i tasks from the P(T ) distribution -where T i is a task composed of a training set (D T rain T i , also known as support set) and a test set (D T est T i , also known as query set) -to train a meta-model that can generalize well to all tasks used in the training process. The trained meta-model can then be fine-tuned to a task T ′ , also sampled from the P(T ) distribution, that was not seen in the meta-model training [Bansal et al 2021]. When dealing with supervised learning tasks using meta-learning, we create the P(T ) distribution based on a fixed set of tasks, subsampled from all classes [Vinyals et al 2016].…”
Section: Dataset Creationmentioning
confidence: 99%
“…The objective of assuming this distribution is to sample T i tasks from the P(T ) distribution -where T i is a task composed of a training set (D T rain T i , also known as support set) and a test set (D T est T i , also known as query set) -to train a meta-model that can generalize well to all tasks used in the training process. The trained meta-model can then be fine-tuned to a task T ′ , also sampled from the P(T ) distribution, that was not seen in the meta-model training [Bansal et al 2021]. When dealing with supervised learning tasks using meta-learning, we create the P(T ) distribution based on a fixed set of tasks, subsampled from all classes [Vinyals et al 2016].…”
Section: Dataset Creationmentioning
confidence: 99%
“…Inspired from the self-supervised learning, Bansal et al (2020b) generates a large number of cloze tasks, which can be considered as multi-class classification tasks but obtained without labeling effort, to augment the meta-training tasks. Bansal et al (2021) further explores the influence of unsupervised task distribution and creates task distributions that are inductive to better meta-training efficacy. The self-supervised generated tasks improve the performance on a wide range of different meta-testing tasks which are classification problems (Bansal et al, 2020b), and it even performs comparably with supervised meta-learning methods on FewRel 2.0 benchmark (Gao et al, 2019b) on 5-shot evaluation (Bansal et al, 2021).…”
Section: Task Augmentationmentioning
confidence: 99%
“…Bansal et al (2021) further explores the influence of unsupervised task distribution and creates task distributions that are inductive to better meta-training efficacy. The self-supervised generated tasks improve the performance on a wide range of different meta-testing tasks which are classification problems (Bansal et al, 2020b), and it even performs comparably with supervised meta-learning methods on FewRel 2.0 benchmark (Gao et al, 2019b) on 5-shot evaluation (Bansal et al, 2021).…”
Section: Task Augmentationmentioning
confidence: 99%