2004
DOI: 10.1109/tip.2004.834664
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Learning of a Finite Mixture Model Based on the Dirichlet Distribution and Its Application

Abstract: This paper presents an unsupervised algorithm for learning a finite mixture model from multivariate data. This mixture model is based on the Dirichlet distribution, which offers high flexibility for modeling data. The proposed approach for estimating the parameters of a Dirichlet mixture is based on the maximum likelihood (ML) and Fisher scoring methods. Experimental results are presented for the following applications: estimation of artificial histograms, summarization of image databases for efficient retriev… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
70
0

Year Published

2005
2005
2021
2021

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 167 publications
(70 citation statements)
references
References 31 publications
0
70
0
Order By: Relevance
“…In particular, model-based clustering using finite mixtures of distributions was discarded because of the low number of observations (42 experts) and the high dimensionality of the JPDs (121 values). We discussed the possible use of finite mixtures of Dirichlet distributions [44] as the most straightforward model for clustering probability distributions. However, the Dirichlet distribution has some constraints, e.g., its covariance matrix is strictly negative so it cannot model positive correlations between variables.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In particular, model-based clustering using finite mixtures of distributions was discarded because of the low number of observations (42 experts) and the high dimensionality of the JPDs (121 values). We discussed the possible use of finite mixtures of Dirichlet distributions [44] as the most straightforward model for clustering probability distributions. However, the Dirichlet distribution has some constraints, e.g., its covariance matrix is strictly negative so it cannot model positive correlations between variables.…”
Section: Discussionmentioning
confidence: 99%
“…Also, it is able to formally address the problem of model selection (finding an appropriate number of clusters). Since each of our observations is aJPD, the Dirichlet distribution [44] could be a suitable choice of a probability density function for each component. However, the low number of observations (N e = 42) over the number of variables (r = 121) ruled out the use of this approach, because it is difficult to obtain accurate estimators of a finite mixture model with so few data.…”
Section: Clustering Of Joint Probability Distributionsmentioning
confidence: 99%
“…The key issue that the mixture models face is the selection of the component density function, which is estimated by the parametric method with a certain parametric model that fits the distribution of the observed objects. Several models have been chosen, including the t-distribution mixture model by Peer and McLachlan [20], the noncentral t-distribution mixture models by Tsung and Jack [21], the hybrid models based on Dirichlet by Bouguila et al [22], and the hybrid models with Gaussian distribution in broad sense by Fan [23]. These models are usually limited in some acquainted distribution for simplifying the problem to be analyzed, resulting in a great difference between basic assumption of the parametric model and real physical model, i.e., model mismatch.…”
Section: Introductionmentioning
confidence: 99%
“…This distribution is the univariate case of the Dirichlet distribution which has proven to have high flexibility to model data (Bouguila et al 2004). Indeed, the capacity of an univariate distribution to provide an accurate fit to data depends on its shape.…”
Section: Introductionmentioning
confidence: 99%