2019
DOI: 10.1007/s10462-019-09784-7
|View full text |Cite
|
Sign up to set email alerts
|

A survey of regularization strategies for deep models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
38
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 117 publications
(61 citation statements)
references
References 28 publications
0
38
0
Order By: Relevance
“…the mean of outputs of ensemble method) is likely to be accurate. In realising the ensemble method, we utilise the dropout technique [21], which is widely used in deep learning models. The dropout technique is commonly used to solve the overfitting problem, typically observed in training a deep learning model.…”
Section: Dropout-based Ensemble Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…the mean of outputs of ensemble method) is likely to be accurate. In realising the ensemble method, we utilise the dropout technique [21], which is widely used in deep learning models. The dropout technique is commonly used to solve the overfitting problem, typically observed in training a deep learning model.…”
Section: Dropout-based Ensemble Methodsmentioning
confidence: 99%
“…if a standard deviation of the distribution is small, an uncertainty for that forecast is low and that forecast is likely to be accurate). We realise the ensemble method by using the dropout technique [21], which is widely used in deep learning forecasting models. Unlike a typical deep learning process that applies the dropout technique only for training a model, we adopt the dropout technique also for a test and that approach allows us to have multiple models and corresponding outputs for the same input.…”
Section: Introductionmentioning
confidence: 99%
“…Since excessive increase in model complexity may also result in overfitting, several regularization techniques can be used to improve model generalizability, such as L 1 and L 2 regularization, batchnormalization, dropout, early stopping, and data augmentation techniques. These techniques can be combined to take advantage of the complementary effects of different approaches, as detailed in a comprehensive overview [53] of the most frequently adopted regularization techniques and of their effects on DL model performance.…”
Section: Deep Learning Modelsmentioning
confidence: 99%
“…When more training leads to an improvement in performance on the training dataset but an otherwise worsening of the performance on the validation dataset, this is a sign that overfitting is occurring which can be typically visualised by plotting so-called loss curves over training time. Overfitting may be prevented by increasing the training dataset’s diversity using, for instance, data augmentation 44 , 45 or using strategies such as reducing the model complexity, adding regularisation (L1, L2) or early stopping during training 46 . DL tools dedicated to training would enormously benefit from these features as these simplify the assessment and potential improvement on model optimisation for the user.…”
Section: Choosing a DL Toolmentioning
confidence: 99%