Difference between revisions of "Posterior Regularization for Expectation Maximization"
From Cohen Courses
Jump to navigationJump to search| Line 15: | Line 15: | ||
The M-step is defined as: | The M-step is defined as: | ||
| + | |||
| + | <math> | ||
| + | \theta^{t+1} = | ||
| + | </math> | ||
Revision as of 17:26, 29 September 2011
Summary
This is a method to impose contraints on posteriors in the Expectation Maximization algorithm, allowing a finer-level control over these posteriors.
Method Description
For a given set x of observed data, a set of latent data z and a set of parameters , the Expectation Maximization algorithm can be viewed as the alternation between two maximization steps of the function .
The E-step is defined as:
where is the Kullback-Leibler divergence given by
The M-step is defined as: