2020
DOI: 10.1007/s42979-020-0071-3
|View full text |Cite
|
Sign up to set email alerts
|

Asymptotic Bayesian Generalization Error in Latent Dirichlet Allocation and Stochastic Matrix Factorization

Abstract: We analyze the asymptotic behavior of the Bayesian generalization error in the topic model. Through a theoretical analysis of the maximum pole of the zeta function (real log canonical threshold) of the topic model, we obtain an upper bound of the Bayesian generalization error and the free energy in the topic model and stochastic matrix factorization (SMF; it can be regarded as a restriction of the non-negative matrix factorization). We show that the generalization error in the topic model and SMF becomes small… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3
1

Relationship

3
1

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 45 publications
0
3
0
Order By: Relevance
“…Concrete values of them were found in the matrix factorization by Aoyagi and Watanabe (2005), the Poisson mixture by Sato and Watanabe (2019), the latent Dirichlet allocation by Hayashi (2021), and many statistical models and learning machines such as Yamazaki and Watanabe (2003); Yamazaki (2016); Yamazaki and Kaji (2013); Zwiernik (2011); Watanabe (2009). They are useful in the design of Markov chain Monte Calro by Nagata and Watanabe (2008) and the singular Bayesian information criterion by Drton and Plummer (2017).…”
Section: Asymptotic Generalization Lossmentioning
confidence: 99%
“…Concrete values of them were found in the matrix factorization by Aoyagi and Watanabe (2005), the Poisson mixture by Sato and Watanabe (2019), the latent Dirichlet allocation by Hayashi (2021), and many statistical models and learning machines such as Yamazaki and Watanabe (2003); Yamazaki (2016); Yamazaki and Kaji (2013); Zwiernik (2011); Watanabe (2009). They are useful in the design of Markov chain Monte Calro by Nagata and Watanabe (2008) and the singular Bayesian information criterion by Drton and Plummer (2017).…”
Section: Asymptotic Generalization Lossmentioning
confidence: 99%
“…According to our prior researches [9,7], several distributions satisfy that the RLCT of NMF is equal to the absolute of the maximum pole of the following zeta function Specifically, when elements of the data matrix follow normal distribution, Poisson distribution, exponential distribution, and Bernoulli distribution, the behavior of the free energy and the generalization error can be describe using RLCT defined by the zeta function . In these previous studies, the prior distribution was limited to positive and bounded, but when proving that the true distribution and the KL information amount of the probabilistic model have the same RLCT as the square norm error of the matrix, the prior distribution is arbitrary.…”
Section: Robustness On Probability Distributionsmentioning
confidence: 99%
“…Determining RLCTs is important for estimating the sufficient sample size, constructing procedures of learning, and selecting models. In fact, RLCTs of many singular models have been studied: mixture models [63,65,47,37,60], Boltzmann machines [67,4,5], non-negative matrix factorization [22,21,18], latent class analysis [13], latent Dirichlet allocation [23,19], naive Bayesian networks [46], Bayesian networks [64],…”
Section: Introductionmentioning
confidence: 99%