2024
DOI: 10.1109/tnnls.2022.3210840
|View full text |Cite
|
Sign up to set email alerts
|

Diminishing Batch Normalization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(1 citation statement)
references
References 16 publications
0
1
0
Order By: Relevance
“…It was beneficial to feature selection to prevent the occurrence of a particularly large neuron weight to reduce overfitting. At the same time, a Batch normalization layer (BN) 36 was added between the full connection layer and the activation function, and the data was divided into small batches for random gradient descent. During the forward transmission, each layer was standardized so that the output value fell into the sensitive area of the input value, and finally, the input of each layer of the neural network keeps the same distribution.…”
Section: Journal Of Chemical Information and Modelingmentioning
confidence: 99%
“…It was beneficial to feature selection to prevent the occurrence of a particularly large neuron weight to reduce overfitting. At the same time, a Batch normalization layer (BN) 36 was added between the full connection layer and the activation function, and the data was divided into small batches for random gradient descent. During the forward transmission, each layer was standardized so that the output value fell into the sensitive area of the input value, and finally, the input of each layer of the neural network keeps the same distribution.…”
Section: Journal Of Chemical Information and Modelingmentioning
confidence: 99%