2012
DOI: 10.1080/10618600.2012.672115
|View full text |Cite
|
Sign up to set email alerts
|

Monotonically Overrelaxed EM Algorithms

Abstract: We explore the idea of overrelaxation for accelerating the expectation-maximization (EM) algorithm, focusing on preserving its simplicity and monotonic convergence properties. It is shown that in many cases a trivial modification in the M-step results in an algorithm that maintains monotonic increase in the log-likelihood, but can have an appreciably faster convergence rate, especially when EM is very slow. The method is applicable to more general fixed point algorithms. Its simplicity and effectiveness are il… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2018
2018
2025
2025

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 34 publications
0
6
0
Order By: Relevance
“…It replaces the M‐step by conducting one iteration of Newton's method. Alternative approaches, such as surrogate functions (Lange et al ., 2000) and overrelaxed EM algorithm (Yu, 2012), have also been introduced in the literature. Pan and Shen (2007) introduced ℓ 1 ‐penalty to the mean parameters for mixture of univariate normal models.…”
Section: Methodsmentioning
confidence: 99%
“…It replaces the M‐step by conducting one iteration of Newton's method. Alternative approaches, such as surrogate functions (Lange et al ., 2000) and overrelaxed EM algorithm (Yu, 2012), have also been introduced in the literature. Pan and Shen (2007) introduced ℓ 1 ‐penalty to the mean parameters for mixture of univariate normal models.…”
Section: Methodsmentioning
confidence: 99%
“…The order-one AA scheme ( 8) is similar but not equivalent to successive overrelaxation (SOR) used in iterative methods for solving large linear systems (e.g., Young (1971)), and it is also similar to the monotonic overrelaxed EM algorithm detailed in Yu (2012). Order-one AA also resembles the STEM procedures described in Varadhan & Roland (2008) which are schemes that utilize Steffensen-type methods to accelerate EM.…”
Section: Anderson Acceleration When M =mentioning
confidence: 99%
“…Second, if the data require many machines for storage, then extensive communication among all the machines further increases the time of each iteration; therefore, EM (Dempster et al 1977) and the family of EM-type algorithms, such as ECM, ECME, AECM, PXEM, and DECME (Meng & Rubin 1993, Liu & Rubin 1994, Meng & van Dyk 1997, Liu et al 1998, He & Liu 2012, are inefficient in massive data settings simply due to the time consuming E step or possibly due to the communication cost. The same is also true for EM extensions that modify the M step by borrowing ideas from optimization (Lange 1995, Jamshidian & Jennrich 1997, Neal & Hinton 1998, Salakhutdinov & Roweis 2003, Varadhan & Roland 2008, Yu 2012.…”
Section: Introductionmentioning
confidence: 92%