2021
DOI: 10.1016/j.neunet.2020.10.009
|View full text |Cite
|
Sign up to set email alerts
|

AutoTune: Automatically Tuning Convolutional Neural Networks for Improved Transfer Learning

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 31 publications
(16 citation statements)
references
References 30 publications
0
14
0
Order By: Relevance
“…( 106 ) proposed a pre-training-based model for problems associated with insufficient data, which effectively reduces the overfitting risks. Several researchers worked with only the first few layers of a pre-trained CNN and then retrained the later ones in a new optimization task ( 107 ). Thus, transfer learning can first train a rough approximation model for a given task and serve as a basis for modeling a novel task.…”
Section: Deep Learning-based Radiomicsmentioning
confidence: 99%
“…( 106 ) proposed a pre-training-based model for problems associated with insufficient data, which effectively reduces the overfitting risks. Several researchers worked with only the first few layers of a pre-trained CNN and then retrained the later ones in a new optimization task ( 107 ). Thus, transfer learning can first train a rough approximation model for a given task and serve as a basis for modeling a novel task.…”
Section: Deep Learning-based Radiomicsmentioning
confidence: 99%
“…The hyperparameters that were optimized were the number of neurons in the FC layers (all neurons are connected to all the neurons in the next layer), weight decay, learning rate, batch size (number of examples used to estimate the gradient), and the number of epochs (number of cycles through the full training data). The number of nodes is an integer between 4 and 128 [32], the weight decay and the learning rate are sampled from a log-uniform distribution with bounds from 10 −10 to 10 −3 and from 10 −5 to 10 −1 [32], respectively, and the batch size is an integer in {4, 8, 16, 32, 64, 128} [35]. With these parameter ranges, the loss-function value stopped improving soon after 10 epochs.…”
Section: Bce(ymentioning
confidence: 99%
“…For this reason, we tried 10, 20, and 30 epochs during the optimization. We used 100 optimization rounds [32] including 20 fully-random start-up rounds [35] to determine the optimal parameter values. The best average accuracy between the two outputs was 0.972 on the validation set.…”
Section: Bce(ymentioning
confidence: 99%
“…An Auto-tuning PID controller architecture with neural networks is shown schematically in Figure 5 . This architecture uses a feedback control unit and a Recurrent Network with Supervised Learning and Delayed Structure to adjust the parameters of the PID controller (how and why the combination was selected can be seen in [ 33 ]). When system parameters vary, the neural network can timely correct the PID controller parameters and properly choose parameters for the input layer based on the problem and system architecture.…”
Section: Introductionmentioning
confidence: 99%