2020
DOI: 10.48550/arxiv.2010.06772
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Scaling Hamiltonian Monte Carlo Inference for Bayesian Neural Networks with Symmetric Splitting

Abstract: Hamiltonian Monte Carlo (HMC) is a Markov chain Monte Carlo (MCMC) approach that exhibits favourable exploration properties in high-dimensional models such as neural networks. Unfortunately, HMC has limited use in large-data regimes and little work has explored suitable approaches that aim to preserve the entire Hamiltonian. In our work, we introduce a new symmetric integration scheme for split HMC that does not rely on stochastic gradients. We show that our new formulation is more efficient than previous appr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(17 citation statements)
references
References 20 publications
0
17
0
Order By: Relevance
“…Since the classic work of Neal (1996), there have been a few recent attempts at using full-batch HMC in BNNs (e.g. ; Cobb & Jalaian, 2020;Wenzel et al, 2020). These studies tend to use relatively short trajectory lengths (generally not considering a number of leapfrog steps greater than 100), and tend to focus on relatively small datasets and networks.…”
Section: Related Workmentioning
confidence: 99%
“…Since the classic work of Neal (1996), there have been a few recent attempts at using full-batch HMC in BNNs (e.g. ; Cobb & Jalaian, 2020;Wenzel et al, 2020). These studies tend to use relatively short trajectory lengths (generally not considering a number of leapfrog steps greater than 100), and tend to focus on relatively small datasets and networks.…”
Section: Related Workmentioning
confidence: 99%
“…Over the last decade, with improving computational resources and advanced MCMC strategies, advanced proposal distributions incorporating gradients have been applied, with Langevin and Hamiltonian MCMC [32,71]. However, only in last five years has there been progress in area of Bayesian neural networks with Hamiltonian MCMC [72], and graphic processing unit (GPU) implementation to enhance computation [73]. Recent work in area where Langevin MCMC methods have been used for neural networks include the use of parallel tempering MCMC for simple neural networks for pattern classifica-tion and time series prediction problems [36].…”
Section: Bayesian Deep Learningmentioning
confidence: 99%
“…The limitations of MCMC has been addressed with better computational resources and advanced proposal distributions, incorporating gradients [29,62]. Hamiltonian MCMC sampling methods have been used for Bayesian neural networks [63] with enhanced computation strategies [64]. Langevin MCMC methods have been used in implementation of Bayesian neural networks for pattern classification and time series prediction problems [65].…”
Section: Bayesian Deep Learningmentioning
confidence: 99%