2005
DOI: 10.1109/tpami.2005.162
|View full text |Cite
|
Sign up to set email alerts
|

Genetic-based EM algorithm for learning Gaussian mixture models

Abstract: We propose a genetic-based expectation-maximization (GA-EM) algorithm for learning Gaussian mixture models from multivariate data. This algorithm is capable of selecting the number of components of the model using the minimum description length (MDL) criterion. Our approach benefits from the properties of Genetic algorithms (GA) and the EM algorithm by combination of both into a single procedure. The population-based stochastic search of the GA explores the search space more thoroughly than the EM method. Ther… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
137
1
10

Year Published

2007
2007
2021
2021

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 231 publications
(160 citation statements)
references
References 14 publications
0
137
1
10
Order By: Relevance
“…The preprocessing (i) consists of a shading correction and a contrast optimization method resulting in a "'clean image"'. The shape segmentation (ii) is executed on the clean image by computing the magnitude of the gradient followed by a fragmentation of the gradient feature space with the Gaussian Mixture Model [15,16] instead of using a static threshold. The result is a probability map describing the probability for the membership to one of the classes (background, embryo).…”
Section: Methodsmentioning
confidence: 99%
“…The preprocessing (i) consists of a shading correction and a contrast optimization method resulting in a "'clean image"'. The shape segmentation (ii) is executed on the clean image by computing the magnitude of the gradient followed by a fragmentation of the gradient feature space with the Gaussian Mixture Model [15,16] instead of using a static threshold. The result is a probability map describing the probability for the membership to one of the classes (background, embryo).…”
Section: Methodsmentioning
confidence: 99%
“…Another random initialization method [10,30] uses mixing proportions equal to 1=K, selects random feature vectors as component means, and employs a spherical covariance matrix equal to 0:1r 2 I, where I is the identity matrix and…”
Section: Related Researchmentioning
confidence: 99%
“…In yet another approach to the GMM parameter estimation problem, the EM algorithm is combined with some global optimization method, e.g.. an evolutionary algorithm [1,30] or a particle swarm optimizer [7]. However, global optimizers have high computational demands and this approach is limited to moderately sized datasets.…”
Section: Related Researchmentioning
confidence: 99%
See 1 more Smart Citation
“…The genetic algorithm [16,17] then creates a population of solutions and applies genetic operators such as mutation and crossover to evolve the solutions in order to find the best one(s).…”
Section: New Approach For Image Segmentationmentioning
confidence: 99%