Seventh IEEE International Conference on Data Mining Workshops (ICDMW 2007) 2007
DOI: 10.1109/icdmw.2007.33
|View full text |Cite
|
Sign up to set email alerts
|

Parallelized Variational EM for Latent Dirichlet Allocation: An Experimental Evaluation of Speed and Scalability

Abstract: Statistical topic models such as the Latent Dirichlet Allocation (LDA)

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2009
2009
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(28 citation statements)
references
References 1 publication
0
28
0
Order By: Relevance
“…Yuan et al (2015) also presents an efficient constant time sampling algorithm for building big topic models. Variational inference can easily be parallelized (Nallapati et al, 2007;Zhai et al, 2012), but has high latency, which has been addressed by performing online updates (Hoffman et al, 2010) and taking stochastic gradients estimated by MCMC inference (Mimno et al, 2012). In this paper, we only focus on single-processor learning, but existing parallelization techniques (Newman et al, 2009) are applicable to our model.…”
Section: Related Workmentioning
confidence: 99%
“…Yuan et al (2015) also presents an efficient constant time sampling algorithm for building big topic models. Variational inference can easily be parallelized (Nallapati et al, 2007;Zhai et al, 2012), but has high latency, which has been addressed by performing online updates (Hoffman et al, 2010) and taking stochastic gradients estimated by MCMC inference (Mimno et al, 2012). In this paper, we only focus on single-processor learning, but existing parallelization techniques (Newman et al, 2009) are applicable to our model.…”
Section: Related Workmentioning
confidence: 99%
“…[47] proposes parallel version of algorithms based on variational EM algorithm for LDA. Two settings of implementations are considered, one is in a multiprocessor architecture and the other is in a distributed environment.…”
Section: Parallel Learning Algorithmsmentioning
confidence: 99%
“…Among these three inference methods, VB inference and collapsed Gibbs sampling have already been parallelized by using PC clusters [8][11] [10]. However, both methods divide a given dataset into smaller subsets and process the subsets in parallel.…”
Section: Collapsed Variational Bayesian Inferencementioning
confidence: 99%