2019
DOI: 10.1016/j.neucom.2019.05.008
|View full text |Cite
|
Sign up to set email alerts
|

Rademacher dropout: An adaptive dropout for deep neural network via optimizing generalization gap

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 23 publications
(9 citation statements)
references
References 10 publications
0
9
0
Order By: Relevance
“…Modified linear unit (ReLU) is used as the activation function, and L2 regularization is introduced to avoid network over-fitting. A dropout layer (Wang et al, 2019) (with random inactivation rate 0.2) is added between the two stacked convolutional layers in the residual module to randomly discard some neurons during training, in order to prevent over-fitting and enhance generalization performance of the network.…”
Section: Residual Modulementioning
confidence: 99%
“…Modified linear unit (ReLU) is used as the activation function, and L2 regularization is introduced to avoid network over-fitting. A dropout layer (Wang et al, 2019) (with random inactivation rate 0.2) is added between the two stacked convolutional layers in the residual module to randomly discard some neurons during training, in order to prevent over-fitting and enhance generalization performance of the network.…”
Section: Residual Modulementioning
confidence: 99%
“…It is understood that as the brain learns stimuli over a period of time, it shows stronger functional connections while having a decrease in randomness within the networks [76]. In contrast, it has been seen in artificial neural networks that random dropouts can benefit learning, and improve the robustness of the networks [77]- [79]. The amount of randomness can have a major impact on the learning ability of the networks.…”
Section: B Randomnessmentioning
confidence: 99%
“…In feature learning and cognition of large-scale multi-state mixed iris data, the statistical learning under the ideal steady state iris training set can ensure intuitive and reliable feature summary. However, over-fitting phenomenon will occur [15]. The cognitive learning of set labels cannot fully evaluate the iris acquisition state in advance, but the initial concept is dependent on the original tags.…”
Section: Cognitive Learning [3]mentioning
confidence: 99%