2019
DOI: 10.1007/s10994-019-05825-y
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference

Abstract: Gradient-based Monte Carlo sampling algorithms, like Langevin dynamics and Hamiltonian Monte Carlo, are important methods for Bayesian inference. In large-scale settings, full-gradients are not affordable and thus stochastic gradients evaluated on mini-batches are used as a replacement. In order to reduce the high variance of noisy stochastic gradients, Dubey et al. [2016] applied the standard variance reduction technique on stochastic gradient Langevin dynamics and obtained both theoretical and experimental i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

1
30
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(39 citation statements)
references
References 8 publications
1
30
0
Order By: Relevance
“…Sampling from a Bayesian posterior distribution lies at the core of many modern machine learning tasks, such as topic modelling [Gan et al, 2015], reinforcement learning [Liu et al, 2017], and Bayesian neural networks [Hernández-Lobato and Adams, 2015]. Particle based Variational Inference (ParVI) methods have recently drawn great attention due to their empirical success in approximating the target posterior distribution [Liu and Wang, 2016;Liu et al, 2017;Feng et al, 2017;Liu and Zhu, 2018]. Typically, these methods update a finite set of interacting particles deterministically to approximately simulate infinite-particle gradient flows on the Wasserstein space P 2 (X ).…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…Sampling from a Bayesian posterior distribution lies at the core of many modern machine learning tasks, such as topic modelling [Gan et al, 2015], reinforcement learning [Liu et al, 2017], and Bayesian neural networks [Hernández-Lobato and Adams, 2015]. Particle based Variational Inference (ParVI) methods have recently drawn great attention due to their empirical success in approximating the target posterior distribution [Liu and Wang, 2016;Liu et al, 2017;Feng et al, 2017;Liu and Zhu, 2018]. Typically, these methods update a finite set of interacting particles deterministically to approximately simulate infinite-particle gradient flows on the Wasserstein space P 2 (X ).…”
Section: Introductionmentioning
confidence: 99%
“…One representative method of this type is the Stein Variational Gradient Descent (SVGD) method [Liu and Wang, 2016], which updates the particles according to a gradient flow described by the Vlasov equation Braun and Hepp, 1977]. Subsequently, by exploiting the Riemannian structure of the Wasserstein space P 2 (X ), [Liu et al, 2019] proposed a Nesterov's-acceleration variant of SVGD called SVGD Wasserstein Nesterov's method (SVGD-WNes).…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations