Structured Prediction Cascades

From Cohen Courses
Jump to navigationJump to search

This method as proposed by Weiss et al, AISTATS 2010

This page is reserved for a write up by Dan Howarth


Citation

Structured Prediction Cascades. David Weiss and Ben Taskar. International Conference on Artificial Intelligence and Statistics (AISTATS), May 2010.

Online version

[1]

Summary

In many structured prediction models an increase in model complexity comes at a high computational cost. For example, the complexity of a HMM grows exponentially with the order of the model. This work introduces a method for learning increasingly complex models while continually pruning the possible output space. This is done by "weeding" out the incorrect output states early on.

Previous methods to solve the problem of model complexity were commonly approximate search methods or heuristic pruning techniques. Structured prediction cascades are different however because they explicitly learn the error/computation trade off for each increase in model complexity.

In this work structured prediction cascades are applied to handwriting recognition and POS tagging.

Brief description of the method

The linear hypothesis class considered is of the form: Failed to parse (MathML with SVG or PNG fallback (recommended for modern browsers and accessibility tools): Invalid response ("Math extension cannot connect to Restbase.") from server "https://wikimedia.org/api/rest_v1/":): {\displaystyle h_w(x) = \argmax_{y \in \mathcal{Y}} \sum_{c \in \mathcal{C}} w^T f_c(x,y_c)}

At each level of the cascade the method is given the set of possible clique assignments as input, and each level then filters this set and passes the filtered set as input to the next level.

is the set of possible output assignments to clique

Let the following be defined for the 'th model:

, the set of maximal cliques

, the set of valid output assignments for clique

Before pruning, , or the set of assignments to the cliques in level that each contain as a subset a valid assignment in the 'th level

Any <math>y_c \in \mathcal{V}^i_c is then pruned if their max-marginal score is less than a threshold.

Experimental Result

Related papers