2021
DOI: 10.24012/dumf.1051352
|View full text |Cite
|
Sign up to set email alerts
|

Effect on model performance of regularization methods

Abstract: Artificial Neural Networks with numerous parameters are tremendously powerful machine learning systems. Nonetheless, overfitting is a crucial problem in such networks. Maximizing the model accuracy and minimizing the amount of loss is significant in reducing in-class differences and maintaining sensitivity to these differences. In this study, the effects of overfitting for different model architectures with the Wine dataset were investigated by Dropout, AlfaDropout, GausianDropout, Batch normalization, Layer n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…The final layer maps all components of the final feature map into two classes which need to be estimated by a softmax activation function. In the U-Net models used, Batch Normalization [39] and Dropout [40] were used together to prevent overfitting at each layer. Dropout is based on the principle of ignoring some randomly selected neurons during training.…”
Section: U-net Architecturementioning
confidence: 99%
“…The final layer maps all components of the final feature map into two classes which need to be estimated by a softmax activation function. In the U-Net models used, Batch Normalization [39] and Dropout [40] were used together to prevent overfitting at each layer. Dropout is based on the principle of ignoring some randomly selected neurons during training.…”
Section: U-net Architecturementioning
confidence: 99%
“…Also, SAE and DBN are unsupervised learning algorithms and do not directly use informations of class while learning features [21]. Convolutional neural networks (CNN) are another deep learning method used for HSIC [22]. CNN has a strong feature learning ability that can ensure more distinctive features for higher quality HSIC [23].…”
mentioning
confidence: 99%