ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021
DOI: 10.1109/icassp39728.2021.9413783
|View full text |Cite
|
Sign up to set email alerts
|

Self-Supervised Learning for Few-Shot Image Classification

Abstract: Few-shot image classification aims to classify unseen classes with limited labelled samples. Recent works benefit from the metalearning process with episodic tasks and can fast adapt to class from training to testing. Due to the limited number of samples for each task, the initial embedding network for meta-learning becomes an essential component and can largely affect the performance in practice. To this end, most of the existing methods highly rely on the efficient embedding network. Due to the limited label… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
37
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 65 publications
(37 citation statements)
references
References 18 publications
0
37
0
Order By: Relevance
“…During pre-training phase, we train the encoder on unsupervised Image900-SSL [8] that contains all images from ImageNet1K except miniImageNet. In addition, the classes in Image900-SSL [8] are distinct from the classes present in CUB.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…During pre-training phase, we train the encoder on unsupervised Image900-SSL [8] that contains all images from ImageNet1K except miniImageNet. In addition, the classes in Image900-SSL [8] are distinct from the classes present in CUB.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…Researchers have developed algorithms to improve the generalization of few-shot learner. Among them, the meta-learning [3,4,5,6,7,8] and fine-tuning methods [9] achieve excellent performance. Notably, both methods described above use Convolutional Neural Network (CNN) encoders, while Vision Transformer (ViT) [10] generalize better than CNN under multiple distribution shifts, which is demonstrated in previous work [11].…”
Section: Introductionmentioning
confidence: 99%
“…PT+MAP 17 and LaplacianShot 18 function similarly, however, both propose alternative strategies for distance metrics when considering query and support points. AmdimNet 19 and S2M2 20 , alternatively, leverage self-supervised techniques in order to generate a stronger embedding-space mapping for input data.…”
Section: Transductive and Self-supervised Approaches To Few-shot Lear...mentioning
confidence: 99%
“…Technique Backbone Preprocessing Extra Training Data AmdimNet 19 Self-supervised Metric AmdimNet No Yes EPNet 16 Transductive Metric WRN28-10 No Yes SimpleCNAPS 14 Metric ResNet18 No Yes PT+MAP 17 Metric WRN28-10 Yes No LaplacianShot 18 Metric WRN28-10 No No S2M2R 20 Self-supervised Metric WRN28-10 Yes No Reptile 13 Optimization CONV4 No No MAML 12 Optimization CONV4 No No ProtoNet 15 Metric CONV4 No No Table 1. An overview of the differing details between the models trained and tested.…”
Section: Model Evaluation Table Model Namementioning
confidence: 99%
“…From this basic learning setting, many extensions have been proposed to improve the performance of metric learning methods. Some of these works focus on pre-training the embedding network [2], others introduce task attention modules [3,11,23], whereas other try to optimize the embeddings [10] and yet others try to use a variety of loss functions [23].…”
Section: Introductionmentioning
confidence: 99%