2022
DOI: 10.48550/arxiv.2207.03859
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Variational Inference of overparameterized Bayesian Neural Networks: a theoretical and empirical study

Abstract: This paper studies the Variational Inference (VI) used for training Bayesian Neural Networks (BNN) in the overparameterized regime, i.e., when the number of neurons tends to infinity. More specifically, we consider overparameterized two-layer BNN and point out a critical issue in the mean-field VI training. This problem arises from the decomposition of the lower bound on the evidence (ELBO) into two terms: one corresponding to the likelihood function of the model and the second to the Kullback-Leibler (KL) div… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 27 publications
(32 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?