Difference between revisions of "Daume ICML 2009"

From Cohen Courses
Jump to navigationJump to search
Line 34: Line 34:
 
* More iterations will cause the output to be closer to the learned classifier and further from the "optimal" policy (which arbitrarily produced the classifications).
 
* More iterations will cause the output to be closer to the learned classifier and further from the "optimal" policy (which arbitrarily produced the classifications).
  
The paper also shows that, given a certain set of definitions, this approach is pretty much equivalent to the [[UsesMethod::Expectation-maximization algorithm|EM algorithm]]
+
The paper also shows that, given a certain set of definitions, this approach is pretty much equivalent to the [[UsesMethod::Expectation-maximization algorithm|EM algorithm]]. It is also noteworthy that the unsupervised algorithm no longer has all of the theoretical guarantees of the supervised algorithm.
  
 
== Experiments ==
 
== Experiments ==
Line 41: Line 41:
  
 
=== Synthetic Experiments ===
 
=== Synthetic Experiments ===
 +
 +
The first experiments involve generating synthetic data from a first order HMM and a second order HMM. The "truth" is produced by using the generating HMM to predict the sequence. The methods compared are: [[UsesMethod::Expectation-maximization algorithm|EM]], [[UsesMethod::SEARN|SEARN]] with [[UsesMethod::HMM|HMM]] features and using [[UsesMethod::Naive Bayes|naive Bayes]], and [[UsesMethod::SEARN|SEARN]] with a [[UsesMethod::Logistic Regression|logistic regression]] classifier.
 +
 +
{| class="wikitable" border="1"
 +
|-
 +
! Model
 +
! States
 +
! Truth (Error Rate)
 +
! EM (Error Rate)
 +
! SEARN-NB (Error Rate)
 +
! SEARN-LR (Error Rate)
 +
|-
 +
| 1st Order HMM
 +
| K = 2
 +
| <math>0.227 \pm0.107</math>
 +
| <math>0.275 \pm0.128</math>
 +
| <math>0.287 \pm0.138</math>
 +
| <math>0.276 \pm0.095</math>
 +
|-
 +
| 1st Order HMM
 +
| K = 5
 +
| <math>0.687 \pm0.043</math>
 +
| <math>0.678 \pm0.026</math>
 +
| <math>0.688 \pm0.025</math>
 +
| <math>0.672 \pm0.022</math>
 +
|-
 +
| 1st Order HMM
 +
| K = 10
 +
| <math>0.806 \pm0.035</math>
 +
| <math>0.762 \pm0.021</math>
 +
| <math>0.771 \pm0.019</math>
 +
| <math>0.755 \pm0.019</math>
 +
|-
 +
| 2nd Order HMM
 +
| K = 2
 +
| <math>0.294 \pm0.072</math>
 +
| <math>0.396 \pm0.057</math>
 +
| <math>0.408 \pm0.056</math>
 +
| <math>0.271 \pm0.057</math>
 +
|-
 +
| 2nd Order HMM
 +
| K = 5
 +
| <math>0.651 \pm0.068</math>
 +
| <math>0.695 \pm0.027</math>
 +
| <math>0.710 \pm0.016</math>
 +
| <math>0.633 \pm0.018</math>
 +
|-
 +
| 2nd Order HMM
 +
| K = 10
 +
| <math>0.815 \pm0.032</math>
 +
| <math>0.764 \pm0.021</math>
 +
| <math>0.771 \pm0.015</math>
 +
| <math>0.705 \pm0.019</math>
 +
|}
 +
 +
The results from the first order HMM are uninteresting, as they are all comparable (i.e. within the standard deviation of the optimal algorithm).
 +
 +
Of the second order HMM results, the best performer is always the [[UsesMethod::SEARN]] with [[UsesMethod::Logistic Regression]]. This is due to the richer set of features that it contains.

Revision as of 23:57, 29 September 2011

In progress by Francis Keith

Citation

"Unsupervised Search-based Structured Prediction", Hal Daume III, ICML 2009

Online Version

An online version of the paper can be found here [1]

Summary

This paper details a methodology for unsupervised SEARN. It compares the results to other methods, first on synthetic data, and then on other methods for unsupervised dependency parsing.

Algorithm

The basic SEARN algorithm is described (see the SEARN wiki article for more background).

In the supervised form, the algorithm uses a sample space of pairs, where is the input and is the true output. In the unsupervised case, we must account for the fact that the algorithm must be run on an input of simply , with the classifier still producing .

The proposed solution is to essentially predict , and then perform the normal prediction. The loss function will only be dependent on , as, while we will be predicting , we do not have any observed "true" outputs.

Given an input , which is a sequence of length , we consider the true output to be of length , where the first components will be drawn from the possible vocabulary for (that is, the possible outputs), and represent the latent structure (which we will refer to as components are the drawn from the vocabulary for the input, . We can then use SEARN on this input. It is important to note that constructing features for these two parts is different:

  • For , the latent components, they can be based on and the partial latent structure (i.e., at , we can look at )
  • For , the input components, they can be based on and . However, the paper notes that in designing features here, the ideal feature set will be dependent on correctly predicting the latent structure correctly. This makes intuitive sense - if we correctly predict , then we have an "optimal" label set.

In dealing with , the optimal (and thus initial) policy, it already predicts fully the true input, and it can arbitrarily/randomly predict the latent input. This gives us the ability to run the first iteration of SEARN.

The paper describes the unsupervised algorithm as follows:

  • First iteration: Use to classify the sequence of latent and input components. Note that randomly or arbitrarily produces classifications for the latent components, and all of the costs are 0, so this won't even induce any update to the classifier for these components. The important piece is that does predict classifications for the input components, and recall that these predictions will produce input symbols, taken from the true input.
  • Second iteration: Given that due to the update, we no longer are guaranteed to produce zero-cost classifications for the latent component. Given that features for the latent structure are based partially on , they get either high costs if the classifier performs poorly, or low costs if it performs well. As a result, the classifier will begin to predict classifications for the latent component.
  • More iterations will cause the output to be closer to the learned classifier and further from the "optimal" policy (which arbitrarily produced the classifications).

The paper also shows that, given a certain set of definitions, this approach is pretty much equivalent to the EM algorithm. It is also noteworthy that the unsupervised algorithm no longer has all of the theoretical guarantees of the supervised algorithm.

Experiments

They use two experiments in the paper. The first is a synthetic experiment using data sampled from HMMs and comparing it to other methods, while the second is applying the method to dependency parsing and comparing it to other, similar methods.

Synthetic Experiments

The first experiments involve generating synthetic data from a first order HMM and a second order HMM. The "truth" is produced by using the generating HMM to predict the sequence. The methods compared are: EM, SEARN with HMM features and using naive Bayes, and SEARN with a logistic regression classifier.

Model States Truth (Error Rate) EM (Error Rate) SEARN-NB (Error Rate) SEARN-LR (Error Rate)
1st Order HMM K = 2
1st Order HMM K = 5
1st Order HMM K = 10
2nd Order HMM K = 2
2nd Order HMM K = 5
2nd Order HMM K = 10

The results from the first order HMM are uninteresting, as they are all comparable (i.e. within the standard deviation of the optimal algorithm).

Of the second order HMM results, the best performer is always the SEARN with Logistic Regression. This is due to the richer set of features that it contains.