2016 IEEE Tenth International Conference on Semantic Computing (ICSC) 2016
DOI: 10.1109/icsc.2016.34
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Tagging Using Topic Models Exploiting Wikipedia Category Network

Abstract: In this paper we propose a probabilistic topic model that incorporates DBpedia knowledge into the topic model for tagging Web pages and online documents with topics discovered in them. Our method is based on integration of the DBpedia hierarchical category network with statistical topic models where DBpedia categories are considered as topics. We have conducted extensive experiments on two different datasets to demonstrate the effectiveness of our method.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
15
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(15 citation statements)
references
References 18 publications
0
15
0
Order By: Relevance
“…In [24] the authors developed a correlated tag learning (CTL) model for the semi-structured corpora based on the topic model to enable the construction of the correlation graph among tags via a logistic normal participation process. In [25] the authors proposed a probabilistic topic model that incorporates DBpedia knowledge into the topic model for tagging web pages and online documents with topics discovered. The model learns the probability distribution of each category over the words using the statistical topic models considering the prior knowledge from Wikipedia about the words, and their associated probabilities in various categories [25].…”
Section: Latent Topic Decompositionmentioning
confidence: 99%
See 2 more Smart Citations
“…In [24] the authors developed a correlated tag learning (CTL) model for the semi-structured corpora based on the topic model to enable the construction of the correlation graph among tags via a logistic normal participation process. In [25] the authors proposed a probabilistic topic model that incorporates DBpedia knowledge into the topic model for tagging web pages and online documents with topics discovered. The model learns the probability distribution of each category over the words using the statistical topic models considering the prior knowledge from Wikipedia about the words, and their associated probabilities in various categories [25].…”
Section: Latent Topic Decompositionmentioning
confidence: 99%
“…In [25] the authors proposed a probabilistic topic model that incorporates DBpedia knowledge into the topic model for tagging web pages and online documents with topics discovered. The model learns the probability distribution of each category over the words using the statistical topic models considering the prior knowledge from Wikipedia about the words, and their associated probabilities in various categories [25]. In [26] the authors proposed Wikipedia-category-concept mention latent Dirichlet allocation (WCM-LDA), which not only models the relationship between words and topics, but also utilizes the concept and category knowledge of entities to model the semantic relation of entities and topics.…”
Section: Latent Topic Decompositionmentioning
confidence: 99%
See 1 more Smart Citation
“…Secondly, as described above, the tags were a set of semantic topic distributions, which were learned from the plain text, and so the correlations should be modeled from the semantic level, while only considering the co-occurrences was not enough. Allahyari, et al (2016). Proposed a probabilistic topic model that incorporates DBpedia Knowledge into the topic model for tagging Web pages and online documents with topics discovered.…”
Section: Taggingmentioning
confidence: 99%
“…Proposed a probabilistic topic model that incorporates DBpedia Knowledge into the topic model for tagging Web pages and online documents with topics discovered. They learn the probability distribution of each category over the words using the statistical topic models taking into account the prior knowledge from Wikipedia about the words, and their associated probabilities in various categories [2]. There were two main limitations in this approach.…”
Section: Taggingmentioning
confidence: 99%