2020
DOI: 10.1002/mp.14507
|View full text |Cite
|
Sign up to set email alerts
|

Targeted transfer learning to improve performance in small medical physics datasets

Abstract: Purpose To perform an in‐depth evaluation of current state of the art techniques in training neural networks to identify appropriate approaches in small datasets. Method In total, 112,120 frontal‐view X‐ray images from the NIH ChestXray14 dataset were used in our analysis. Two tasks were studied: unbalanced multi‐label classification of 14 diseases, and binary classification of pneumonia vs non‐pneumonia. All datasets were randomly split into training, validation, and testing (70%, 10%, and 20%). Two popular c… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
24
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(34 citation statements)
references
References 34 publications
1
24
0
Order By: Relevance
“…Our results demonstrate that the relative effect of transfer learning is indirectly proportional to sample size. This finding is consistent with machine learning theory 62,63 and implies that the performance of our sclerotic glomeruli model may asymptotically improve with more data. Second, data augmentation was also shown to increase the generalization of our DL results, suggesting artificial augmentations may capture important characteristics of renal pathology 64 .…”
Section: Discussionsupporting
confidence: 89%
“…Our results demonstrate that the relative effect of transfer learning is indirectly proportional to sample size. This finding is consistent with machine learning theory 62,63 and implies that the performance of our sclerotic glomeruli model may asymptotically improve with more data. Second, data augmentation was also shown to increase the generalization of our DL results, suggesting artificial augmentations may capture important characteristics of renal pathology 64 .…”
Section: Discussionsupporting
confidence: 89%
“…AG-Mask Interface 87.1 DL [ 136 ] lr = 0.0001 Adam Softmax 224 224 Images are resized using the bilinear interpolation and are also horizontally flipped CXR-14 ResNet-18, Transfer learning, CAM, Data augmentation 84.94 DL [ 126 ] lr = 0.00005 SGD Sigmoid 1024 1024 Images are resized using the bilinear interpolation and additional techniques such as rotation, zoom and shifting are performed for data augmentation AMC, SNUBH ResNet-50, fine-tuning, CAM, curriculum learning 98.3 DL [ 178 ] lr = 0.0001 Adam ReLu, Sigmoid 128 128 Images are normalized to enhance the contrast. Further, rotation, scaling and flipping is performed for reducing the overfitting CXR-14 Inception-ResNet-v2, dilated ResNet, transfer Learning, cross-validation 90.0 DL [ 44 ] lr = 0.0001 and it is decreased by 10 times after every 3 epochs Adam Softmax, Sigmoid 224 224 Images normalization is performed CXR-14 ResNet-18, DenseNet-121, MSML 84.32 DL [ 140 ] 236 236 Images are resized, cropped, flipped and rotated …”
Section: Classificationmentioning
confidence: 99%
“…All the pre-trained models are generally trained on the ImageNet dataset. Romero et al [ 140 ] have reviewed, evaluated, and compared the state-of-the-art training techniques on the small datasets. Deep CNN model performance has been checked on small datasets for emphysema detection, pneumonia detection, hernia detection, and CXR-14 classification with or without transfer learning.…”
Section: Classificationmentioning
confidence: 99%
“…Considering that most of tasks are correlated, transfer learning can accelerate and optimize the learning efficiency of the model by transferring the pre-trained model parameters to the new model, rather than learning from scratch. For deep learning models, the transfer method applies the pre-training model to a new task by fine-tuning ( 105 ). For the transfer learning method, the source domain and the target domain do not need the same distribution of data, which overcomes the shortcomings of traditional machine learning, and has a great advantage in the case of few samples in the target domain and sufficient samples in related fields.…”
Section: Discussionmentioning
confidence: 99%