2014
DOI: 10.1007/s10994-014-5476-6
|View full text |Cite
|
Sign up to set email alerts
|

Additive regularization of topic models

Abstract: Probabilistic topic modeling of text collections has been recently developed mainly within the framework of graphical models and Bayesian inference. In this paper we introduce an alternative semi-probabilistic approach, which we call additive regularization of topic models (ARTM). Instead of building a purely probabilistic generative model of text we regularize an ill-posed problem of stochastic matrix factorization by maximizing a weighted sum of the log-likelihood and additional criteria. This approach enabl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0
5

Year Published

2017
2017
2019
2019

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 90 publications
(42 citation statements)
references
References 24 publications
0
35
0
5
Order By: Relevance
“…where (a) + def = max(a, 0) [27]; β x , α y and γ z are the elements of the hyperparameter vectors β, α and γ, respectively; and…”
Section: A Learning: Em-algorithm Schemementioning
confidence: 99%
See 1 more Smart Citation
“…where (a) + def = max(a, 0) [27]; β x , α y and γ z are the elements of the hyperparameter vectors β, α and γ, respectively; and…”
Section: A Learning: Em-algorithm Schemementioning
confidence: 99%
“…The corresponding Dirichlet distributions with all used parameters are presented in Figure 4. Note that parameter learning is an ill-posed problem in topic modeling [27]. This means there is no unique solution for parameter estimates.…”
Section: A Setupmentioning
confidence: 99%
“…A number of papers is devoted to the sparseness of the target distributions, e.g., [29]. In [156] different forms of regularisation are presented to overcome the problem of non-uniqueness of the matrix factorisation in topic modeling.…”
Section: Extensions Of Conventional Modelsmentioning
confidence: 99%
“…where (a) + def = max(a, 0) [156]; η w , α z , κ b , and υ b are the elements of the hyperparameter vectors η, α, κ and υ, respectively, and:…”
Section: Expectation-maximisation Learningmentioning
confidence: 99%
“…First, there is a natural way to learn document embeddings. Second, additive regularization of topic models [43] can be used to meet further requirements. In this paper we employ…”
Section: Additive Regularization and Embeddings For Multiple Modalitiesmentioning
confidence: 99%