Topic modelling approaches help scholars to examine the topics discussed in a corpus. Due to the popularity of Twitter, two distinct methods have been proposed to accommodate the brevity of tweets: the tweet pooling method and Twitter LDA. Both of these methods demonstrate a higher performance in producing more interpretable topics than the standard Latent Dirichlet Allocation (LDA) when applied on tweets. However, while various metrics have been proposed to estimate the coherence of the generated topics from tweets, the coherence of the top ranked topics, those that are most likely to be examined by users, has not been investigated. In addition, the effect of the number of generated topics K on the topic coherence scores has not been studied. In this paper, we conduct large-scale experiments using three topic modelling approaches over two Twitter datasets, and apply a state-of-the-art coherence metric to study the coherence of the top ranked topics and how K affects such coherence. Inspired by ranking metrics such as precision at n, we use coherence at n to assess the coherence of a topic model. To verify our results, we conduct a pairwise user study to obtain human preferences over topics. Our findings are threefold: we find evidence that Twitter LDA outperforms both LDA and the tweet pooling method because the top ranked topics it generates have more coherence; we demonstrate that a larger number of topics (K) helps to generate topics with more coherence; and finally, we show that coherence at n is more effective when evaluating the coherence of a topic model than the average coherence score.
[1]
Mark Steyvers,et al.
Finding scientific topics
,
2004,
Proceedings of the National Academy of Sciences of the United States of America.
[2]
Iadh Ounis,et al.
Topic-centric classification of Twitter user's political orientation
,
2015
.
[3]
Timothy Baldwin,et al.
Automatic Evaluation of Topic Coherence
,
2010,
NAACL.
[4]
David Buttler,et al.
Exploring Topic Coherence over Many Models and Many Topics
,
2012,
EMNLP.
[5]
Scott Sanner,et al.
Improving LDA topic models for microblogs via tweet pooling and automatic labeling
,
2013,
SIGIR.
[6]
Michael I. Jordan,et al.
Latent Dirichlet Allocation
,
2001,
J. Mach. Learn. Res..
[7]
Craig MacDonald,et al.
Using Word Embedding to Evaluate the Coherence of Topics from Twitter Data
,
2016,
SIGIR.
[8]
Qiaozhu Mei,et al.
Understanding the Limiting Factors of Topic Modeling via Posterior Contraction Analysis
,
2014,
ICML.
[9]
Hongfei Yan,et al.
Comparing Twitter and Traditional Media Using Topic Models
,
2011,
ECIR.
[10]
Craig MacDonald,et al.
Topics in Tweets: A User Study of Topic Coherence Metrics for Twitter Data
,
2016,
ECIR.
[11]
Thomas L. Griffiths,et al.
Probabilistic Topic Models
,
2007
.
[12]
Brian D. Davison,et al.
Empirical study of topic modeling in Twitter
,
2010,
SOMA '10.