Sutton McCullum ICML 2007: Piecewise pseudolikelihood for efficient CRF training

From Cohen Courses
Jump to navigationJump to search

Citation

Piecewise Pseudolikelihood for Efficient Training of Conditional Random Fields. By Charles Sutton, Andrew McCallum. In ICML, vol. {{{volume}}} ({{{issue}}}), 2007.

Online version

This Paper is available here.

Summary

Discriminative training of graphical models is expensive if the cardinality of the variables is large. Generally pseudo-likelihood reduces the cost of inference, but compromises on accuracy. Piecewise training although is accurate, is expensive in a similar way. The authors try to maximize the pseudo-likelihood on the piecewise model.

Definition of Piecewise Pseudo likelihood

For a single instance ,


Thus, piecewise pseudolikelihood is a sum of local conditional log-probabilities. The local probabilities are defined as


Therefore the optimization function is

where the second term is the standard gaussian prior to prevent over fitting.

Compared to standard piecewise which requires time (where is the maximum cardinality of the label and is the maximum size of a factor), PWPL requires . Compared to pseudolikelihood where each term conditions on the entire Markov blanket, in PWPL, it conditions on its neighbors.

Experiments

The authors propose that PWPL performs better on small datasets whereas pseudo likelihood performs better on large datasets. They verify this by generating data from a second order [HMM] with transition and emission probabilities as a linear combination (). Higher represents more complexity and deviation from the assumption of first-order. For different values of , they generate 1000 sequences of length 25 and 1000 for 150 synthetic generating models. A first-order CRF is trained over these sets.

On an average PWPL performs identically to standard piecewise training. Although pseudolikelihood performs better than PWPL, the result is not statistically significant. However, when the accuracy as a function of training set size is compared, pseudolikelihood converges to a limit higher than PWPL.

Other experiments conducted included POS Tagging over the Penn Treebank, Noun-phrase chunking and Named Entity Recognition over CoNLL'03. The performance in terms of time taken is significantly less in all the three tasks.