Difference between revisions of "Gimpel and Smith, NAACL 2010"

From Cohen Courses
Jump to navigationJump to search
Line 1: Line 1:
 
'''Softmax-Margin CRFs: Training Log-Linear Models with Cost Functions'''
 
'''Softmax-Margin CRFs: Training Log-Linear Models with Cost Functions'''
  
Online: [http://www.cs.cmu.edu/~kgimpel/papers/gimpel+smith.naacl10.pdf]
+
This [[Category:Paper|paper]] can be found at: [http://www.cs.cmu.edu/~kgimpel/papers/gimpel+smith.naacl10.pdf]
  
 
==Citation==
 
==Citation==

Revision as of 17:56, 25 September 2011

Softmax-Margin CRFs: Training Log-Linear Models with Cost Functions

This can be found at: [1]

Citation

Kevin Gimpel and Noah A. Smith. Softmax-margin CRFs: Training log-linear models with loss functions. In Proceedings of the Human Language Technologies Conference of the North American Chapter of the Association for Computational Linguistics, pages 733-736, Los Angeles, California, USA, June 2010.

Summary

The authors want to be able to incorporate a cost function (present in structured SVMs) into standard conditional log-likelihood models. They introduce the softmax-margin objective function that achieves the best of both worlds. Using a NER task, it performs significantly better than a standard conditional loglikelihood model, a max-margin model, and the perceptron, but is indistinguishable from MIRA, risk, and JRB (Jensen risk bound; defined in the paper).

Brief Description of the Softmax-Margin objective function

Consider the objective functions for these four methods. Our

Conditional log likelihood:

Max-margin:

Risk:

Softmax-margin:

Experimental Results

Related Work