2019
DOI: 10.48550/arxiv.1907.12087
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Charting the Right Manifold: Manifold Mixup for Few-shot Learning

Abstract: Few-shot learning algorithms aim to learn model parameters capable of adapting to unseen classes with the help of only a few labeled examples. A recent regularization technique -Manifold Mixup focuses on learning a generalpurpose representation, robust to small changes in the data distribution. Since the goal of few-shot learning is closely linked to robust representation learning, we study Manifold Mixup in this problem setting. Self-supervised learning is another technique that learns semantically meaningful… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(17 citation statements)
references
References 39 publications
(64 reference statements)
0
17
0
Order By: Relevance
“…Some meta learning models need to pretrain on a lager task of N-way K-shot before training on 5-way 5-shot(1-shot), which is called meta pretraining (Snell, Swersky, and Zemel 2017). Moreover, some models use self-supervised pretraining (Mangla et al 2019) or pertrained feature extractor (Lee et al 2019). However, our framework can be meta-trained end-to-end without any method of pretraining.…”
Section: Resultsmentioning
confidence: 99%
“…Some meta learning models need to pretrain on a lager task of N-way K-shot before training on 5-way 5-shot(1-shot), which is called meta pretraining (Snell, Swersky, and Zemel 2017). Moreover, some models use self-supervised pretraining (Mangla et al 2019) or pertrained feature extractor (Lee et al 2019). However, our framework can be meta-trained end-to-end without any method of pretraining.…”
Section: Resultsmentioning
confidence: 99%
“…The Nesterov Momentum optimizer [50] is used with an initial learning rate of 0.01. The total training epochs on the CUB, the miniImageNet and the Kinetics are 57, 40 and 23, and the learning rate is dropped to 10% on (30,40), (30,37), (2,19) epochs respectively. The weight decay is set to be 0.0005.…”
Section: Methodsmentioning
confidence: 99%
“…Self-supervised learning aims at learning from the supervision of the object structure, alleviating the need of supervision from manual labels, and has been researched in the field of unsupervised/semisupervised learning [31,42]. Recently, this mechanism is also applied in FSL by methods such as predicting the rotation [19,37] and predicting the relative position [19]. In this work, inspired by these previous works, we propose to use the self-supervised split loss to learn part-related primitives and alleviate the influence of the semantic gap between known and novel classes.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations