Difference between revisions of "Martins et al 2010"
(8 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
+ | === Citation and Online Link === | ||
+ | |||
A. F. T. Martins, K. Gimpel. N. A. Smith, E. P. Xing, P. M. Q. Aguiar, M. A. T. Figueiredo, 2010. [http://www.cs.cmu.edu/~nasmith/papers/martins+etal.tr10.pdf Aggressive Online Learning of Structured Classifiers]. Technical report CMU-ML-10-109. | A. F. T. Martins, K. Gimpel. N. A. Smith, E. P. Xing, P. M. Q. Aguiar, M. A. T. Figueiredo, 2010. [http://www.cs.cmu.edu/~nasmith/papers/martins+etal.tr10.pdf Aggressive Online Learning of Structured Classifiers]. Technical report CMU-ML-10-109. | ||
+ | |||
+ | === Summary === | ||
+ | |||
+ | This [[Category::paper]] generalizes the loss function of CRFs, structured SVMs, structured perceptron, and Softmax-margin CRFs into a single loss function, and then derives an online learning algorithm that can be used to learn with that more general loss function. For the hinge loss, the learning algorithm reduces to MIRA. | ||
+ | |||
+ | === Method === | ||
+ | |||
+ | The general loss function is: | ||
+ | |||
+ | [[file:Martins et al 2010 Loss Function.png]] | ||
+ | |||
+ | Different choices of <math>\Beta</math> and <math>\gamma</math> correspond to various well known loss functions. They are: | ||
+ | |||
+ | [[file:Martins et al 2010 Parameter Choices.png]] | ||
+ | |||
+ | The function minimized is the loss function with a regularizer: | ||
+ | |||
+ | [[file:Martins et al 2010 Learning Problem.png]] [[file:Martins et al Relarizer.png]] | ||
+ | |||
+ | [[file:Martins et al Regularize Coeff.png]] | ||
+ | |||
+ | The online learning algorithm proposed to minimize this function is called Dual Coordinate Ascent (DCA): | ||
+ | |||
+ | [[file:Martins et al 2010 DCA.png]] | ||
+ | |||
+ | The parameters can be updated using algorithm 2: | ||
+ | |||
+ | [[file:Martins et al 2010 Alg2.png]] | ||
+ | |||
+ | === Experimental Result === | ||
+ | |||
+ | === Related Papers === | ||
+ | MIRA | ||
+ | CRF | ||
+ | Softmax-margin CRFs | ||
In progress by [[User:Jmflanig]] | In progress by [[User:Jmflanig]] |
Latest revision as of 02:08, 11 October 2011
Citation and Online Link
A. F. T. Martins, K. Gimpel. N. A. Smith, E. P. Xing, P. M. Q. Aguiar, M. A. T. Figueiredo, 2010. Aggressive Online Learning of Structured Classifiers. Technical report CMU-ML-10-109.
Summary
This paper generalizes the loss function of CRFs, structured SVMs, structured perceptron, and Softmax-margin CRFs into a single loss function, and then derives an online learning algorithm that can be used to learn with that more general loss function. For the hinge loss, the learning algorithm reduces to MIRA.
Method
The general loss function is:
Different choices of and correspond to various well known loss functions. They are:
The function minimized is the loss function with a regularizer:
The online learning algorithm proposed to minimize this function is called Dual Coordinate Ascent (DCA):
The parameters can be updated using algorithm 2:
Experimental Result
Related Papers
MIRA CRF Softmax-margin CRFs
In progress by User:Jmflanig