2009
DOI: 10.1111/j.1467-9868.2009.00698.x
|View full text |Cite
|
Sign up to set email alerts
|

On-Line Expectation–Maximization Algorithm for latent Data Models

Abstract: We propose a generic on-line (also sometimes called adaptive or recursive) version of the expectation-maximization (EM) algorithm applicable to latent variable models of independent observations. Compared with the algorithm of Titterington, this approach is more directly connected to the usual EM algorithm and does not rely on integration with respect to the complete-data distribution. The resulting algorithm is usually simpler and is shown to achieve convergence to the stationary points of the Kullback-Leible… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

7
426
0
1

Year Published

2010
2010
2022
2022

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 366 publications
(434 citation statements)
references
References 20 publications
7
426
0
1
Order By: Relevance
“…Since our online approach builds upon the EM formalism presented by Cappé et al [4], our notation in the following derivations is closely related to that of the latter. If we write our parameter vector θ as…”
Section: A Preliminariesmentioning
confidence: 99%
See 2 more Smart Citations
“…Since our online approach builds upon the EM formalism presented by Cappé et al [4], our notation in the following derivations is closely related to that of the latter. If we write our parameter vector θ as…”
Section: A Preliminariesmentioning
confidence: 99%
“…Figure 5 shows an example of online parameter estimation using our EM framework, on a data set comprising 5000 samples. Cappé et al [4] resolve issues due to the dependency on the initialization by updating onlyŝ for a number of first iterations. Thus, we use the first 100 observations to build up an estimate ofŝ (100) , before calculating the first parameter estimateθ (101) .…”
Section: Online Estimationmentioning
confidence: 99%
See 1 more Smart Citation
“…The online EM was first introduced in [25]. It was subsequently generalized in [26], where the authors present a proof convergence and a proposition of asymptotic equivalency to natural gradient ascent. Moreover, the algorithm performs simple and efficient updates, making good use of the sufficient statistics of the exponential family, so as to avoid redundant calculations.…”
Section: A Proposed Approachmentioning
confidence: 99%
“…For instance, if the coefficients of the reward functions evolve at each time-step according to a linear model, then its parameters can be estimated through the online expectation maximization algorithm [20]. In the dynamic MABC problem, an accurate model of the dynamics can improve not only the estimation accuracy of the reward function parameters, but also provide valuable information for the action selection process.…”
Section: Estimation Of Time-varying Reward Functionsmentioning
confidence: 99%