2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00501
|View full text |Cite
|
Sign up to set email alerts
|

Iterative Normalization: Beyond Standardization Towards Efficient Whitening

Abstract: Batch Normalization (BN) is ubiquitously employed for accelerating neural network training and improving the generalization capability by performing standardization within mini-batches. Decorrelated Batch Normalization (DBN) further boosts the above effectiveness by whitening. However, DBN relies heavily on either a large batch size, or eigendecomposition that suffers from poor efficiency on GPUs. We propose Iterative Normalization (IterNorm), which employs Newtons iterations for much more efficient whitening,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
148
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 93 publications
(149 citation statements)
references
References 24 publications
1
148
0
Order By: Relevance
“…The whitening procedure, a.k.a decorrelated batch normalization, does not only standardize the feature but also eliminates the data correlation. The decorrelated batch normalization can improve both the optimization efficiency and generalization ability of deep neural networks (Huang et al, 2018;Siarohin et al, 2018;Huang et al, 2019;Pan et al, 2019;Huang et al, 2020;Ermolov et al, 2021).…”
Section: Applicationsmentioning
confidence: 99%
See 3 more Smart Citations
“…The whitening procedure, a.k.a decorrelated batch normalization, does not only standardize the feature but also eliminates the data correlation. The decorrelated batch normalization can improve both the optimization efficiency and generalization ability of deep neural networks (Huang et al, 2018;Siarohin et al, 2018;Huang et al, 2019;Pan et al, 2019;Huang et al, 2020;Ermolov et al, 2021).…”
Section: Applicationsmentioning
confidence: 99%
“…Table 1 summarizes the forward computational complexity. As suggested in Li et al (2018); Huang et al (2019), the iteration times for NS iteration are often set as 5 such that reasonable performances can be achieved. That is, to consume the same complexity as the NS iteration does, our MTP and MPA can match to the power series up to degree 16.…”
Section: Matrix Pad é Approximantsmentioning
confidence: 99%
See 2 more Smart Citations
“…There exists extensive study on optimizing effectiveness via tuning batch size. On the one hand, a small batch size will lead to a high variance of statistics and weaken the training stability (Wu and He 2018;Huang et al 2019a). On the other hand, a large batch size can reduce the estimation noise but it will incur a sharp landscape of loss (Keskar et al 2016) making the optimization problem more challenging.…”
Section: Introductionmentioning
confidence: 99%