2007
DOI: 10.1109/tpami.2007.1095
|View full text |Cite
|
Sign up to set email alerts
|

High-Dimensional Unsupervised Selection and Estimation of a Finite Generalized Dirichlet Mixture Model Based on Minimum Message Length

Abstract: We consider the problem of determining the structure of high-dimensional data, without prior knowledge of the number of clusters. Data are represented by a finite mixture model based on the generalized Dirichlet distribution. The generalized Dirichlet distribution has a more general covariance structure than the Dirichlet distribution and offers high flexibility and ease of use for the approximation of both symmetric and asymmetric distributions. This makes the generalized Dirichlet distribution more practical… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
65
0

Year Published

2011
2011
2020
2020

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 137 publications
(65 citation statements)
references
References 48 publications
0
65
0
Order By: Relevance
“…However, the Dirichlet distribution has some constraints, e.g., its covariance matrix is strictly negative so it cannot model positive correlations between variables. Finite mixtures of generalized Dirichlet distributions [66] overcome some of these constraints. However, the generalized Dirichlet distribution has more parameters than the Dirichlet distribution, so the problem of high-dimensionality combined with few data is even more challenging.…”
Section: Discussionmentioning
confidence: 99%
“…However, the Dirichlet distribution has some constraints, e.g., its covariance matrix is strictly negative so it cannot model positive correlations between variables. Finite mixtures of generalized Dirichlet distributions [66] overcome some of these constraints. However, the generalized Dirichlet distribution has more parameters than the Dirichlet distribution, so the problem of high-dimensionality combined with few data is even more challenging.…”
Section: Discussionmentioning
confidence: 99%
“…0.2.4, p. 537, col. 2] for a discussion of MML hierarchical clustering. As well as the abovementioned multinomial, Gaussian, Poisson and von Mises circular distributions [Wallace and Dowe, 1994b;1996;1997b;, this work -without sequential and spatial clustering (following in the next item) has been extended to other distributions Dowe, 2002a, 2002b;2003a;2003b;Bouguila and Ziou, 2007];…”
Section: Elusive Model Paradox (And Encryption)mentioning
confidence: 99%
“…p(u|z) and p(e|z) denote the likelihood of a user and context to belong respectively to the user's class z. p(r|z, c) is the probability to generate a rating for given user and image classes. We model p(v|c) using the Generalized Dirichlet distribution (GDD) [3][2] which is suitable for non Gaussian data such as images. This distribution has a more general covariance structure and provides multiple shapes.…”
Section: The Visual Content Context Flexible Mixture Modelmentioning
confidence: 99%
“…The Fisher information of θ cl and ξ l can be computed by considering the loglikelihood of each feature taken separately [3]. After the second order derivations of this log-likelihood, we obtain:…”
Section: Model Selection and Parameter Estimation Using MMLmentioning
confidence: 99%
See 1 more Smart Citation