2021
DOI: 10.1016/j.mlwa.2021.100124
|View full text |Cite
|
Sign up to set email alerts
|

A study of the generalizability of self-supervised representations

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(11 citation statements)
references
References 44 publications
0
10
0
Order By: Relevance
“…While both object detection and instance segmentation target localization of arbitrary class objects, 3DHPSE only targets a single class, the human. For inference on arbitrary class objects, learning a wide range of general features unlimited to labels of a dataset could be advantageous in the generalization aspect (Tendle & Hasan, 2021). However, for 3DHPSE, a backbone network is preferred to learn more about human features rather than features of arbitrary objects, given the limited learning capacity.…”
Section: Pre-training On Imagenetmentioning
confidence: 99%
“…While both object detection and instance segmentation target localization of arbitrary class objects, 3DHPSE only targets a single class, the human. For inference on arbitrary class objects, learning a wide range of general features unlimited to labels of a dataset could be advantageous in the generalization aspect (Tendle & Hasan, 2021). However, for 3DHPSE, a backbone network is preferred to learn more about human features rather than features of arbitrary objects, given the limited learning capacity.…”
Section: Pre-training On Imagenetmentioning
confidence: 99%
“…Hence, SSL is not limited to learning only the label-relevant features that help predict the frequent classes, but rather a diverse set of generalizable representations, including both label-relevant and irrelevant features from unlabeled data. Learning during the pretext task also contributes to the representation-invariance property of an SSL model (Tendle & Hasan, 2021), such that it captures the ingrained characteristics of the input distribution, that are generalizable or transferable to downstream tasks. Therefore, SSL methods can generalize to rare classes better than SL approaches.…”
Section: F I G U R E 1 1 Confusion Matrix For Nearest Neighbor Contra...mentioning
confidence: 99%
“…Therefore, SSL methods can generalize to rare classes better than SL approaches. SSL's robustness to class imbalance is thoroughly demonstrated by Liu, Zhang et al (2021), and the generalizability of self-supervised representations is discussed by Tendle and Hasan (2021).…”
Section: F I G U R E 1 1 Confusion Matrix For Nearest Neighbor Contra...mentioning
confidence: 99%
“…While supervised monocular depth estimation currently outperforms self-supervised methods, their performance is converging towards that of supervised ones. Additionally, research has shown that selfsupervised methods are better at generalizing across a variety of environments [26] (e.g., indoor/outdoor, urban/rural scenes). Many works assume the entire 3D world is a rigid scene, thus ignoring objects that move independently.…”
Section: Introductionmentioning
confidence: 99%