2021
DOI: 10.48550/arxiv.2110.09510
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unsupervised Finetuning

Abstract: This paper studies "unsupervised finetuning", the symmetrical problem of the well-known "supervised finetuning". Given a pretrained model and small-scale unlabeled target data, unsupervised finetuning is to adapt the representation pretrained from the source domain to the target domain so that better transfer performance can be obtained. This problem is more challenging than the supervised counterpart, as the low data density in the small-scale target data is not friendly for unsupervised learning, leading to … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 29 publications
0
3
0
Order By: Relevance
“…In contrast to source-free domain adaptation and test time training, our target pretrain does not modify training objective in the source domain nor assume the prediction task to be identical between the source and target domains. A recent related work addressed finetuning on target domain in an unsupervised manner [33]. To avoid large deviations from pretrained model parameters they simultaneously finetune on both target and source domain data.…”
Section: B Domain Adaptationmentioning
confidence: 99%
“…In contrast to source-free domain adaptation and test time training, our target pretrain does not modify training objective in the source domain nor assume the prediction task to be identical between the source and target domains. A recent related work addressed finetuning on target domain in an unsupervised manner [33]. To avoid large deviations from pretrained model parameters they simultaneously finetune on both target and source domain data.…”
Section: B Domain Adaptationmentioning
confidence: 99%
“…Benefiting from its concise form, UniGrad can be easily extended with commonly used data augmentations [3,4,23,29,32,34] to further boost its performance. As a We apply PCA to representations, and count the number of eigenvalues whose cumulative sum first exceeds 90%.…”
Section: Application On Data Augmentationsmentioning
confidence: 99%
“…As a We apply PCA to representations, and count the number of eigenvalues whose cumulative sum first exceeds 90%. demonstration, we show how to apply CutMix [23,32] and multi-crop [3,4] to our method below.…”
Section: Application On Data Augmentationsmentioning
confidence: 99%
“…This approach would allow creation of a customized model without spending time on a distinct calibration stage. Similar techniques have been studied in other machine learning applications [86]- [89].…”
Section: Opportunities To Reduce Calibrationmentioning
confidence: 90%