2023
DOI: 10.24251/hicss.2023.116
|View full text |Cite
|
Sign up to set email alerts
|

A Practical and Empirical Comparison of Three Topic Modeling Methods Using a COVID-19 Corpus: LSA, LDA, and Top2Vec

Ferhat Zengul,
Aysegul Bulut,
Nurettin Oner
et al.

Abstract: This study was prepared as a practical guide for researchers interested in using topic modeling methodologies. This study is specially designed for those with difficulty determining which methodology to use. Many topic modeling methods have been developed since the 1980s namely, latent semantic indexing or analysis (LSI/LSA), , probabilistic LSI/LSA (pLSI/pLSA), naïve Bayes, the Author-Recipient-Topic (ART), Latent Dirichlet Allocation (LDA), Topic Over Time (TOT), Dynamic Topic Models (DTM), Word2Vec, Top2Vec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(1 citation statement)
references
References 30 publications
0
1
0
Order By: Relevance
“…The algebraic topic model, such as Latent Semantic Allocation (LSA), was developed in the 1990s [2], represents the corpus as a document term matrix (DTM). Zengul et al [12]state that LSA and topic modeling are among the most commonly employed methods. LSA is a natural language processing approach that examines associations between text-based terms and documents, assuming that words with similar meanings will occur in similar contexts.…”
Section: Introductionmentioning
confidence: 99%
“…The algebraic topic model, such as Latent Semantic Allocation (LSA), was developed in the 1990s [2], represents the corpus as a document term matrix (DTM). Zengul et al [12]state that LSA and topic modeling are among the most commonly employed methods. LSA is a natural language processing approach that examines associations between text-based terms and documents, assuming that words with similar meanings will occur in similar contexts.…”
Section: Introductionmentioning
confidence: 99%