Difference between revisions of "Smith and Eisner 2005:Contrastive Estimation: Training Log-Linear Models on Unlabeled Data"

From Cohen Courses
Jump to navigationJump to search
Line 11: Line 11:
 
This is an interesting [[Category::paper]] that presents an unsupervised [[UsesMethod::Contrastive Estimation]] method for [[UsesMethod::Conditional Random Fields]] and other [[AddressesProblem::Log-Linear Models]], which can be easily applied to estimation problems in [[AddressesProblem::Part of Speech Tagging]], [[AddressesProblem::Named Entity Recognition]] and [[AddressesProblem::Semantic Role Labeling]]. When applying this technique to POS tagging, the observed results outperforms [[UsesMethod::Expectation Maximization]], and is robust when the dictionary quality is poor.
 
This is an interesting [[Category::paper]] that presents an unsupervised [[UsesMethod::Contrastive Estimation]] method for [[UsesMethod::Conditional Random Fields]] and other [[AddressesProblem::Log-Linear Models]], which can be easily applied to estimation problems in [[AddressesProblem::Part of Speech Tagging]], [[AddressesProblem::Named Entity Recognition]] and [[AddressesProblem::Semantic Role Labeling]]. When applying this technique to POS tagging, the observed results outperforms [[UsesMethod::Expectation Maximization]], and is robust when the dictionary quality is poor.
  
== Brief description of the method ==
+
== Brief Descriptions of the Method ==
  
 
== The Datasets ==  
 
== The Datasets ==  

Revision as of 23:33, 29 September 2011

Citation

Smith, Noah A. and Jason Eisner (2005). Contrastive estimation: Training log-linear models on unlabeled data. Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL), pages 354-362, Ann Arbor, Michigan, June.

Online version

Smith and Eisner 2005

Summary

This is an interesting paper that presents an unsupervised Contrastive Estimation method for Conditional Random Fields and other Log-Linear Models, which can be easily applied to estimation problems in Part of Speech Tagging, Named Entity Recognition and Semantic Role Labeling. When applying this technique to POS tagging, the observed results outperforms Expectation Maximization, and is robust when the dictionary quality is poor.

Brief Descriptions of the Method

The Datasets

Experimental Results

The authors conduct two major experiments: first, they compare the contrastive estimation approach with existing EM approach on POS tagging with unlabeled data; secondly, the author has also experimented with the robustness of this approach by removing knowledge from the lexicon, and adding features. To analyze the experiment results precisely, we take the original figures and tables in the experiment section from the Smith and Eisner 2005 paper.

Comparison with EM

Ce exp1.png

In the above figure, it is clear that fully supervised methods are still the state-of-the-art methods for POS tagging when labeled data are present. However, when comparing to EM in an unsupervised setting, the LENGTH model obtains an Viterbi accuracy of 78.9%, which is significantly higher than the EM baseline (60.9%) on the 96K-word dataset. DEL1WORD and DEL1SUBSEQ are even worse than the EM on larger datasets. The DELORTR1 and TRAN1 models also achieve good results.

Evaluation on Robustness in the Unsupervised Settings

Ce exp2.png

In Table 3, it shows the experiment results when tagging dictionary is tweaked. It is observed that when removing knowledge from the lexicon, all models are worse than before. Among them, the LENGTH model suffers the most. DELORTRAN1 and TRANS1 do not manipulate emission weights for OOV, so it does not suffer so badly. The most interesting finding of this experiment is that when adding the spelling features, the DELORTRANS1 and TRANS1 model almost completely recover the original accuracy, though LENGTH does not fully recovered.

Related papers

From a structured prediction perspective, this paper presents an interesting contrastive estimation approach that can be compared with many existing estimation techniques, for example, joint likelihood maximization in HMMs, conditional likelihood estimation and sum of conditional likelihoods. Secondly, this paper is also in line with some other numerical optimization approach that optimizes a convex objective function. Moreover, from the empirical evaluation standpoint, the proposed unsupervised approach might not be able to outperform the standard supervised POS tagging, but it can be applied to some sequential modeling tasks where labeled data are not abundantly available, for example, Named Entity Tagging, Parts-of-speech Tagging, and Constituent Parsing for resource-poor languages. Below shows some of the related papers to this work.