Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval 2005
DOI: 10.1145/1076034.1076148
|View full text |Cite
|
Sign up to set email alerts
|

Relation between PLSA and NMF and implications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
131
0
2

Year Published

2008
2008
2021
2021

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 200 publications
(144 citation statements)
references
References 6 publications
3
131
0
2
Order By: Relevance
“…We choose LDA for extracting latent topic labels among radiology report documents because LDA is shown to be more flexible yet learns more coherent topics over large sets of documents [43]. Furthermore, pLSA can be regarded as a special case of LDA [13] and NMF as a semi-equivalent model of pLSA [12,10]. LDA offers a hierarchy of extracted topics and the number of topics can be chosen by evaluating each model's perplexity score (Equation 1), which is a common way to measure how well a probabilistic model generalizes by evaluating the log-likelihood of the model on a held-out test set.…”
Section: Document Topic Learning With Latent Dirichlet Allocationmentioning
confidence: 99%
“…We choose LDA for extracting latent topic labels among radiology report documents because LDA is shown to be more flexible yet learns more coherent topics over large sets of documents [43]. Furthermore, pLSA can be regarded as a special case of LDA [13] and NMF as a semi-equivalent model of pLSA [12,10]. LDA offers a hierarchy of extracted topics and the number of topics can be chosen by evaluating each model's perplexity score (Equation 1), which is a common way to measure how well a probabilistic model generalizes by evaluating the log-likelihood of the model on a held-out test set.…”
Section: Document Topic Learning With Latent Dirichlet Allocationmentioning
confidence: 99%
“…Given this connection, one can also establish a relation between LDA and certain settings of low-rank matrix factorization. Specifically, Gaussier and Goutte [8] and Ding et al [4] have noted that pLSA correspond to specific instances of the problem of non-negative matrix factorization. pLSA can thus be reduced to a low-rank matrix factorization problem.…”
Section: Islda ≈ Random Matrix Factorizationmentioning
confidence: 99%
“…( [2] refers to its algorithm as probabilistic latent semantic analysis (PLSA). Under proper normalization and for the KL objective function used in this paper, NMF and PLSA are numerically equivalent [3], so the results in [2] are equally relevant to NMF or PLSA.) NMF works well for separating sounds when the building blocks for different sources are sufficiently distinct.…”
Section: Introductionmentioning
confidence: 99%