2017
DOI: 10.1007/s11042-017-5145-4
|View full text |Cite
|
Sign up to set email alerts
|

Conceptualization topic modeling

Abstract: Recently, topic modeling has been widely used to discover the abstract topics in text corpora. Most of the existing topic models are based on the assumption of three-layer hierarchical Bayesian structure, i.e. each document is modeled as a probability distribution over topics, and each topic is a probability distribution over words. However, the assumption is not optimal. Intuitively, it's more reasonable to assume that each topic is a probability distribution over concepts, and then each concept is a probabil… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 17 publications
(3 citation statements)
references
References 24 publications
(10 reference statements)
0
3
0
Order By: Relevance
“…Although the proposed method takes the dependency information among words and semantics associations in the text into account, it abstracts away from the influence of time on the sentiment and topic. The following study will consider the use of the existing inference algorithms such as the collapse of Gibbs sampling 45,46 and variational Bayesian 51–57 to detect dynamic topic and sentiment. We also consider developing a more flexible and efficient inference algorithm to accommodate the training requirements of the large‐scale corpus.…”
Section: Discussionmentioning
confidence: 99%
“…Although the proposed method takes the dependency information among words and semantics associations in the text into account, it abstracts away from the influence of time on the sentiment and topic. The following study will consider the use of the existing inference algorithms such as the collapse of Gibbs sampling 45,46 and variational Bayesian 51–57 to detect dynamic topic and sentiment. We also consider developing a more flexible and efficient inference algorithm to accommodate the training requirements of the large‐scale corpus.…”
Section: Discussionmentioning
confidence: 99%
“…Applications of Probase include semantic search [94][95][96], understanding Web tables [97], question answering [90], short text understanding [98,99] and classification [100], topic modelling [101], open directory based text classification [102], and learning entity and concept representation [103].…”
Section: Characteristicsmentioning
confidence: 99%
“…This is also why topic models based on word embeddings are, in this context, a superior solution to algorithms that use concept databases, such asTang et al (2018), or algorithms that model concepts simply as latent variables, asEl-Arini, Fox, and Guestrin (2012).…”
mentioning
confidence: 99%