2021
DOI: 10.48550/arxiv.2110.07161
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Attention-Aware Hierarchical Topic Model

Abstract: Neural topic models (NTMs) apply deep neural networks to topic modelling. Despite their success, NTMs generally ignore two important aspects: (1) only document-level word count information is utilized for the training, while more fine-grained sentence-level information is ignored, and (2) external semantic knowledge regarding documents, sentences and words are not exploited for the training. To address these issues, we propose a variational autoencoder (VAE) NTM model that jointly reconstructs the sentence and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
4

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(10 citation statements)
references
References 22 publications
0
6
4
Order By: Relevance
“…However, not all authors apply document classification or clustering algorithms before performing topic discovery. The results obtained in this paper are higher than the results obtained by the authors [8][9][10][12][13][14]. On the other hand, authors such as [14] obtain higher coherence values for the 20-Newsgroup corpus.…”
Section: Resultscontrasting
confidence: 84%
See 2 more Smart Citations
“…However, not all authors apply document classification or clustering algorithms before performing topic discovery. The results obtained in this paper are higher than the results obtained by the authors [8][9][10][12][13][14]. On the other hand, authors such as [14] obtain higher coherence values for the 20-Newsgroup corpus.…”
Section: Resultscontrasting
confidence: 84%
“…The results obtained in this paper are higher than the results obtained by the authors [8][9][10][12][13][14]. On the other hand, authors such as [14] obtain higher coherence values for the 20-Newsgroup corpus. The corpus Reuters [23,24] obtain higher coherence values than the results obtained in this work.…”
Section: Resultscontrasting
confidence: 84%
See 1 more Smart Citation
“…Topic modeling has always been a catalyst for other research areas in Natural Language Process (NLP) (Panwar et al 2020;Jin et al 2021;Srivastava and Sutton 2016). A classic statistical topic model is Latent Dirichlet Allocation (LDA), which is based on Gibbs sampling to extract topics from documents (Blei, Ng, and Jordan 2003).…”
Section: Related Work Topic Modelmentioning
confidence: 99%
“…Bahrainian S A et al [21] proposed a new light-weight Self-Supervised Neural Topic Model (SNTM) that learns a rich context by learning a topic representation jointly from three co-occurring words and a document that the triplet originates from. Jin Y et al [22] proposed a variational autoencoder (VAE) NTM model that jointly reconstructs the sentence and document word counts using combinations of bag-of-words (BoW) topical embeddings and pre-trained semantic embeddings. Zhao H et al [23] proposed to learn the topic distribution of a document by directly minimising its OT distance to the document's word distributions.…”
Section: Neural Topic Modelingmentioning
confidence: 99%