2019
DOI: 10.1007/978-3-030-22796-8_28
|View full text |Cite
|
Sign up to set email alerts
|

Uncertainty Estimation via Stochastic Batch Normalization

Abstract: In this work, we investigate Batch Normalization technique and propose its probabilistic interpretation. We propose a probabilistic model and show that Batch Normalization maximazes the lower bound of its marginalized log-likelihood. Then, according to the new probabilistic model, we design an algorithm which acts consistently during train and test. However, inference becomes computationally inefficient. To reduce memory and computational cost, we propose Stochastic Batch Normalization -an efficient approximat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
32
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 33 publications
(34 citation statements)
references
References 4 publications
(7 reference statements)
1
32
0
Order By: Relevance
“…We denote the normalization operation as F (·) and the normalized output asx = F (X B ; x). For a certain x, X B can be viewed as a random variable [2,46].x is thus a random variable which shows the stochasticity. It's interesting to explore the statistical momentum of x to measure the magnitude of the stochasticity.…”
Section: Stochastic Normalization Disturbancementioning
confidence: 99%
“…We denote the normalization operation as F (·) and the normalized output asx = F (X B ; x). For a certain x, X B can be viewed as a random variable [2,46].x is thus a random variable which shows the stochasticity. It's interesting to explore the statistical momentum of x to measure the magnitude of the stochasticity.…”
Section: Stochastic Normalization Disturbancementioning
confidence: 99%
“…Another work (Atanov et al 2018) interpret the mean and variance of mini-batch statistics used in batch normalization (Ioffe and Szegedy 2015) as random variables since they depend on stochastic shuffling of training examples into mini-batches during training. Thus, the neural network with batch normalization layers can be viewed as a probabilistic model during training.…”
Section: Related Workmentioning
confidence: 99%
“…We apply stochastic batch normalization (Atanov et al 2018) to obtain uncertainty estimation on a deep neural network trained for detecting diabetic retinopathy. Unlike previous works that demonstrate out-of-dataset detection by artificially splitting a dataset by classes (CIFAR5) or generating new images by rotation (notMINIST), we observe that domain shift in real-world dataset is more subtle.…”
Section: Introductionmentioning
confidence: 99%
“…Concurrently to this work, the authors of [2] have proposed a similar model for BN stochasticity and demonstrated that the distributions of U and V can be used at test time for improving the test data likelihoods and out-of-domain uncertainties. However, they did not explore using this model during the learning.…”
Section: Model Of Bn Stochasticitymentioning
confidence: 96%
“…There are several closely related works concurrent with this submission [20,25,2,15]. Work [20] argues that BN improves generalization because it leads to a smoother objective function, the authors of [15] study the question why BN is often found incompatible with dropout, and works [25,2] observe that randomness in batch normalization can be linked to optimizing a lower bound on the expected data likelihood [2] and to variational Bayesian learning [25]. However, these works focus on estimating the uncertainty of outputs in models that have been already trained using BN.…”
Section: Related Workmentioning
confidence: 99%