Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval 2016
DOI: 10.1145/2911451.2914731
|View full text |Cite
|
Sign up to set email alerts
|

Examining the Coherence of the Top Ranked Tweet Topics

Abstract: Topic modelling approaches help scholars to examine the topics discussed in a corpus. Due to the popularity of Twitter, two distinct methods have been proposed to accommodate the brevity of tweets: the tweet pooling method and Twitter LDA. Both of these methods demonstrate a higher performance in producing more interpretable topics than the standard Latent Dirichlet Allocation (LDA) when applied on tweets. However, while various metrics have been proposed to estimate the coherence of the generated topics from … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
10
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
2
1

Relationship

3
5

Authors

Journals

citations
Cited by 17 publications
(14 citation statements)
references
References 12 publications
2
10
0
Order By: Relevance
“…In general, the coherence values increased along with the number of topics as aforementioned. The result is in line with that in the data mining literature [49]. Conversely, the mean value of spatial autocorrelation decreased.…”
Section: Inferring Activity Types and Entangled Urban Functionssupporting
confidence: 91%
See 1 more Smart Citation
“…In general, the coherence values increased along with the number of topics as aforementioned. The result is in line with that in the data mining literature [49]. Conversely, the mean value of spatial autocorrelation decreased.…”
Section: Inferring Activity Types and Entangled Urban Functionssupporting
confidence: 91%
“…In general, the higher the coherence value, the better the quality of the topics. We implemented a grid search of best topic models and ended up with the same conclusion as that in [49]. Increasing the number of topics (T) leads to higher coherence values.…”
Section: Indicators For Select the Number Of Topicsmentioning
confidence: 99%
“…We show the coherence of the topic models extracted from the two candidate communities in Appendix Figure B1. The coherence results are consistent with Fang, MacDonald, Ounis, and Habel (2016a): the average coherence of a topic model decreases when the number of topics increases; however, the increasing line of c@10/20/30 in Figure B1 indicates that the top-ranked topics in a topic model are much easier to understand as K increases. Among proClinton topic models, we found the coherence (c@10/20/30) of topics becomes stable when K reaches 70, and for proTrump, when K reaches 60.…”
Section: Discussionsupporting
confidence: 74%
“…Meanwhile, we also examine the top 2/7 2 most coherent topics in a model for more effective coherence evaluation, i.e. C@2 & C@7 metrics, following to [24].…”
Section: Methodsmentioning
confidence: 99%