Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1640
|View full text |Cite
|
Sign up to set email alerts
|

Topic Modeling with Wasserstein Autoencoders

Abstract: We propose a novel neural topic model in the Wasserstein autoencoders (WAE) framework. Unlike existing variational autoencoder based models, we directly enforce Dirichlet prior on the latent document-topic vectors. We exploit the structure of the latent space and apply a suitable kernel in minimizing the Maximum Mean Discrepancy (MMD) to perform distribution matching. We discover that MMD performs much better than the Generative Adversarial Network (GAN) in matching high dimensional Dirichlet distribution. We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
75
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 83 publications
(77 citation statements)
references
References 12 publications
2
75
0
Order By: Relevance
“…NVI: the VAE‐based approach introduced in Miao et al (2016). Was‐A: The Wasserstein Autoencoder model presented in Nan et al (2019). AVITM: This technique is published in Srivastava and Sutton (2017), using VAE to approximate the LDA process. AATM:The model uses the attention‐based autoencoder technique for topic modelling in (Tian & Fang, 2019).…”
Section: Experiments and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…NVI: the VAE‐based approach introduced in Miao et al (2016). Was‐A: The Wasserstein Autoencoder model presented in Nan et al (2019). AVITM: This technique is published in Srivastava and Sutton (2017), using VAE to approximate the LDA process. AATM:The model uses the attention‐based autoencoder technique for topic modelling in (Tian & Fang, 2019).…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…This method is based on probability and word distribution, making it difficult to process short text. A similar approach is presented in Nan, Ding, Nallapati, and Xiang (2019), in which the deep neural network is used to model the Dirichlet process in LDA. Work reported in Zhu, Feng, and Li (2018) introduces an approach that is similar to the clustering process, when an enhanced graph‐based VAE is proposed to capture varied learned words from large corpus and apply it to topic modelling problem.…”
Section: Related Workmentioning
confidence: 99%
“…We use the preprocessed 20Newsgroups of (Srivastava and Sutton, 2017), and preprocessed Grolier and NYTimes of (Wang et al, 2019a). We compare the performance of our model with LDA (Blei et al, 2003), NVDM (Miao et al, 2016), ProdLDA (Srivastava and Sutton, 2017), GraphBTM (Zhu et al, 2018), ATM (Wang et al, 2019a) and W-LDA (Nan et al, 2019) using topic coherence measures (Röder et al, 2015). To quantify the understandability of the extracted topics, a topic coherence measure aggregates the relatedness scores of the topic words (topweighted words) of each topic, where the word relatedness scores are estimated based on word co-occurrence statistics on a large external corpus.…”
Section: Methodsmentioning
confidence: 99%
“…We impose a Dirichlet prior, the conjugate prior of the multinomial distribution, to the latent topic distributions. Following W-LDA (Nan et al, 2019), we achieve this goal by minimizing the Maximum Mean Discrepancy (MMD) (Gretton et al, 2012) between the distribution QẐ of inferred topic distributionsẑ and the Dirichlet prior P Z from which we draw multinomial noises z:…”
Section: Training Objectivementioning
confidence: 99%
See 1 more Smart Citation