Difference between revisions of "Posterior Regularization for Expectation Maximization"

From Cohen Courses
Jump to navigationJump to search
Line 1: Line 1:
 
== Summary ==
 
== Summary ==
  
This is a [[Category::method]] to impose contraints on posteriors in the [[AddressesProblem::Expectation Maximization]] algorithm, allowing a finer-level control over these posteriors.
+
The [[Expectation Maximization]] algorithm is a method for finding the maximum likelihood estimates for the parameters in a statistical model. During the E-step of this algorithm, posterior probabilities are calculated for the latent data by fixing the parameters.
 +
 
 +
In many fields, prior knowledge about the posterior probabilities are known and can be applied to the model to improve the statistical model, yet the method to include such knowledge is not always clear.
 +
 
 +
Posterior Regularization is a [[Category::method]] used to impose contraints on posteriors in the [[AddressesProblem::Expectation Maximization]] algorithm, allowing a finer-level control over these posteriors, which allows prior knowledge to be defined.
  
 
== Method Description ==
 
== Method Description ==

Revision as of 23:07, 29 September 2011

Summary

The Expectation Maximization algorithm is a method for finding the maximum likelihood estimates for the parameters in a statistical model. During the E-step of this algorithm, posterior probabilities are calculated for the latent data by fixing the parameters.

In many fields, prior knowledge about the posterior probabilities are known and can be applied to the model to improve the statistical model, yet the method to include such knowledge is not always clear.

Posterior Regularization is a method used to impose contraints on posteriors in the Expectation Maximization algorithm, allowing a finer-level control over these posteriors, which allows prior knowledge to be defined.

Method Description

For a given set of observed data, a set of latent data and a set of parameters , the Expectation Maximization algorithm can be viewed as the alternation between two maximization steps of the function , by marginalizing different free variables.

The E-step is defined as:

where is the Kullback-Leibler divergence given by , and q(z|x) is an arbitrary probability distribution over the latent variable z.

The M-step is defined as:

The goal of this method is to define a way to constrains over posteriors.