2007
DOI: 10.1016/j.csda.2006.10.028
|View full text |Cite
|
Sign up to set email alerts
|

Mean-field variational approximate Bayesian inference for latent variable models

Abstract: The ill-posed nature of missing variable models offers a challenging testing ground for new computational techniques. This is the case for the mean-field variational Bayesian inference (Jaakkola, 2001; Beal and Ghahramani, 2002; Beal, 2003). In this note, we illustrate the behavior of this approach in the setting of the Bayesian probit model. We show that the mean-field variational method always underestimates the posterior variance and, that, for small sample sizes, the mean-field variational approximation to… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
30
0

Year Published

2011
2011
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(33 citation statements)
references
References 9 publications
3
30
0
Order By: Relevance
“…Further trials (not presented here) on the data of Section V, on synthetic truth data and on more complex/highly sparse benchmark data (e.g., the Forensic Glass set [3]) suggest that the proposed mean field VB approximations can perform badly with very sparse data sets and are instead most reliable for intermediate and large sample sizes (i.e., relative to the number of parameters to be estimated). This agrees with the findings of previous empirical and theoretical studies on the asymptotic properties of mean field VB approximations in other related models [15], [18], [35], which show that VB estimates are generally biased with respect to the true posterior statistics (as an unavoidable consequence of 'deliberate model misspecification' via the mean field assumption). While such biases can be quite significant at small sample sizes, they generally become much less significant as the proportion of observed to unobserved variables in the training set increases (or equivalently, as the number of training data increases when the number of hidden parameters can be fixed, as in the models studied here).…”
Section: A Performance Considerationssupporting
confidence: 91%
See 1 more Smart Citation
“…Further trials (not presented here) on the data of Section V, on synthetic truth data and on more complex/highly sparse benchmark data (e.g., the Forensic Glass set [3]) suggest that the proposed mean field VB approximations can perform badly with very sparse data sets and are instead most reliable for intermediate and large sample sizes (i.e., relative to the number of parameters to be estimated). This agrees with the findings of previous empirical and theoretical studies on the asymptotic properties of mean field VB approximations in other related models [15], [18], [35], which show that VB estimates are generally biased with respect to the true posterior statistics (as an unavoidable consequence of 'deliberate model misspecification' via the mean field assumption). While such biases can be quite significant at small sample sizes, they generally become much less significant as the proportion of observed to unobserved variables in the training set increases (or equivalently, as the number of training data increases when the number of hidden parameters can be fixed, as in the models studied here).…”
Section: A Performance Considerationssupporting
confidence: 91%
“…From (5), the posterior over the hidden variables takes exactly the same form as (15), where the denominator is now given by marginalization of (34) over the variables. Once again, substitution of (35)-(40) into (15) leads to an intractable posterior, as in the MMS case.…”
Section: Variational Bayes Learning For Me Modelsmentioning
confidence: 98%
“…This is clearly seen in the centre graphs where the θ i values where i > p have been "switched off". The results in both examples are similar in that the variational posterior noise distribution is more compact than the actual, a tendency reported by a number of authors (for example [23], [24], [25]), the estimated θ values are similar to the actuals, with a tendency to be underestimated, and the reconstructions of the data are good.…”
Section: A Synthesising Datasupporting
confidence: 71%
“…Various algorithms for VAEs (Kingma & Welling, ; Maddison et al, ; Jang et al, ) can be used to optimize this objective. In these VAE algorithms, mean‐field family (Blei & Jordan, ; Consonni & Marin, ) is assumed for the working model Q ( Z | X , θ ), that is, Q ( Z | X , θ ), can be factorized with respect to Z . Moreover, Q ( X | Z , f ) can also be factorized across components of X , which is consistent with our generating scheme that assumes X | Z is mutually independent.…”
Section: A Deep Knockoff Generator Through Latent Variablesmentioning
confidence: 99%