2021
DOI: 10.48550/arxiv.2101.05913
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Supervised Transfer Learning at Scale for Medical Imaging

Abstract: Transfer learning is a standard technique to improve performance on tasks with limited data. However, for medical imaging, the value of transfer learning is less clear [38]. This is likely due to the large domain mismatch between the usual natural-image pre-training (e.g. ImageNet) and medical images. However, recent advances in transfer learning have shown substantial improvements from scale. We investigate whether modern methods can change the fortune of transfer learning for medical imaging. For this, we st… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
31
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
4

Relationship

2
8

Authors

Journals

citations
Cited by 23 publications
(32 citation statements)
references
References 33 publications
1
31
0
Order By: Relevance
“…To study ViT representations, we draw on techniques from neural network representation similarity, which allow the quantitative comparisons of representations within and across neural networks [17,34,26,19]. These techniques have been very successful in providing insights on properties of different vision architectures [29,22,18], representation structure in language models [48,25,47,21], dynamics of training methods [33,24] and domain specific model behavior [27,35,38]. We also apply linear probes in our study, which has been shown to be useful to analyze the learned representations in both vision [1] and text [8,32,45] models.…”
Section: Related Workmentioning
confidence: 99%
“…To study ViT representations, we draw on techniques from neural network representation similarity, which allow the quantitative comparisons of representations within and across neural networks [17,34,26,19]. These techniques have been very successful in providing insights on properties of different vision architectures [29,22,18], representation structure in language models [48,25,47,21], dynamics of training methods [33,24] and domain specific model behavior [27,35,38]. We also apply linear probes in our study, which has been shown to be useful to analyze the learned representations in both vision [1] and text [8,32,45] models.…”
Section: Related Workmentioning
confidence: 99%
“…Large scale transfer learning by pre-training on JFT , Dosovitskiy et al, 2020, Ryoo et al, 2021, Mustafa et al, 2021, Tay et al, 2021a, Puigcerver et al, 2020, Ngiam et al, 2018 or ImageNet21K [Dosovitskiy et al, 2020, Mustafa et al, 2021, Puigcerver et al, 2020 has been done extensively. Mensink et al [2021] considers a two-step transfer chain, where the model is pre-trained on ImageNet, fine-tuned on the source task and then transferred to the target task.…”
Section: Appendixmentioning
confidence: 99%
“…Since model advances are quickly translated to real-world, safety-critical applications (e.g. Mustafa et al 2021), there is an urgent need to re-assess the calibration properties of current state-of-the-art models.…”
Section: Introductionmentioning
confidence: 99%