2022
DOI: 10.1109/access.2022.3163270
|View full text |Cite
|
Sign up to set email alerts
|

Revisiting Bayesian Autoencoders With MCMC

Abstract: Autoencoders gained popularity in the deep learning revolution given their ability to compress data and provide dimensionality reduction. Although prominent deep learning methods have been used to enhance autoencoders, the need to provide robust uncertainty quantification remains a challenge. This has been addressed with variational autoencoders so far. Bayesian inference via Markov Chain Monte Carlo (MCMC) sampling methods have faced limitations; however, recent advances in parallel computing and advanced pro… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 81 publications
0
6
0
Order By: Relevance
“…Furthermore, future research could test other ML models to capture the network structure in SMD. One potential avenue in this context -depending on the dataset dimensions -is the application of other probabilistic, bayesian latent space models, neural networks and autoencoders (e.g., Chandra et al, 2022;Radev et al, 2020;Yong & Brintrup, 2022). Such models provide the opportunity to explore the posterior distributions of opinion estimates, potentially leading to even more accurate predictions of sociopolitical outcomes.…”
Section: Discussionmentioning
confidence: 99%
“…Furthermore, future research could test other ML models to capture the network structure in SMD. One potential avenue in this context -depending on the dataset dimensions -is the application of other probabilistic, bayesian latent space models, neural networks and autoencoders (e.g., Chandra et al, 2022;Radev et al, 2020;Yong & Brintrup, 2022). Such models provide the opportunity to explore the posterior distributions of opinion estimates, potentially leading to even more accurate predictions of sociopolitical outcomes.…”
Section: Discussionmentioning
confidence: 99%
“…In the case of Bayesian neural networks, the prior distribution can be based on the distribution of the weights and biases from similar neural network models. This can be seen as an example of expert knowledge and implemented in previous studies [69], [104]. Another example of expert knowledge is the concept of weight decay [105] regularisation (L2 or Ridge regression [106]) which restricts large weights and can be incorporated when defining the prior distribution (priors) [8], [8], [9].…”
Section: Mcmcmentioning
confidence: 99%
“…The method has shown to be effective for linear models [61] which motivated its use in Bayesian neural networks. In the literature, Langevin MCMC has been very promising for simple and deep neural networks [62], [69], [70]. Hence, we draw the proposed values for the parameters (θ p ) according to a one-step (epoch) gradient as shown in Equation 39.…”
Section: A: Langevin Proposal Distributionmentioning
confidence: 99%
See 1 more Smart Citation
“…To resolve the computational burden arising from the hierarchical Bayesian model, we consolidate the Bayesian sampling procedure via Stochastic Gradient Langevin Dynamics (SGLD) with an adaptive empirical Bayesian variable selection method using Expectation-maximization. Instead of computing the full batch gradient, SGLD evaluates mini-batch gradients with injected random Gaussian noise, which is theoretically valid to generate Langevin-based proposal distribution [18,19]. The mini-batch gradient learning naturally fits into the training of deep neural networks and relaxes the scalability issue at the same time.…”
Section: Introductionmentioning
confidence: 99%