Difference between revisions of "Structured SVMs"

From Cohen Courses
Jump to navigationJump to search
 
(4 intermediate revisions by the same user not shown)
Line 1: Line 1:
Being edited by Rui Correia
+
== The Method and When to Use it==
 
 
== The Method and When to Use it ==
 
  
 
Structured (or Structural) Support Vector Machines (SSVM), as the name states, is a machine learning model that generalizes the [[Support Vector Machines| Support Vector Machine (SVM)]] classifier, allowing training a classifier for structured output.
 
Structured (or Structural) Support Vector Machines (SSVM), as the name states, is a machine learning model that generalizes the [[Support Vector Machines| Support Vector Machine (SVM)]] classifier, allowing training a classifier for structured output.
  
In general, SSVMs perform supervised learning by approximating a mapping <math>h</math>
+
In general, SSVMs perform supervised learning by approximating a mapping <math>f</math>
  
 
<math>
 
<math>
h: X \rightarrow Y
+
f: X \rightarrow Y
 
</math>
 
</math>
  
Line 26: Line 24:
 
</math>
 
</math>
  
where <math>\Delta</math> is an arbitrary function, which measures the distance between to labels and <math>\Psi</math> is a function on samples and labels, which extracts feature vectors.
+
where <math>\Delta</math> is an arbitrary function, which measures the distance between to labels and <math>\Psi</math> is a function on samples and labels, which extracts feature vectors. These two function are to be defined according to the problem at hand
  
  
Line 44: Line 42:
  
 
== Testing ==
 
== Testing ==
At test time, only a sample is known, and a prediction function maps it to a predicted label from the label space . For structured SVMs, given the vector obtained from training, the prediction function is the following.
+
Given a sample, <math>x \in X </math> and a mapping <math>f: X \rightarrow Y</math> one can obtain the correspondent lable. The mapping is defined as
 +
 
 +
 
 +
<math>
 +
f(x) = \arg\max_{y \in Y} w'\Psi(x,y)
 +
</math>
 +
 
 +
where <math>w</math> is the vector during the training phase.
  
Therefore, the maximizer over the label space is the predicted label. Solving for this maximizer is the so called inference problem and similar to making a maximum a-posteriori (MAP) prediction in probabilistic models. Depending on the structure of the function Ψ, solving for the maximizer can be a hard problem.
+
The inference problem of solving for this maximizer over the label space is dependent on the structure of the function <math>\Psi</math> that varies according to the problem.
  
 
== Related Papers ==
 
== Related Papers ==
*[http://svmlight.joachims.org/svm_struct.html| I. Tsochantaridis, T. Hofmann, T. Joachims, and Y. Altun. Support Vector Learning for Interdependent and Structured Output Spaces, ICML, 2004.]
+
*[[Taskar, B. et al, NIPS 2003| Taskar 
et 
al.
 (2003)]]
* Optimization Algorithms
+
*[http://svmlight.joachims.org/svm_struct.html| Tsochantaridis et al. (2004)]
**Taskar 
et 
al.
(2003):

 SMO
 based 
on 
factored 
dual

+
*[[Taskar et al. 2004. Max-margin Parsing| Taskar et al. (2004)]]
**Bartlet
 et 
al. 
(2004) 
and 
Collins 
et 
al. 
(2008):

 exponentiated 
gradient

+
*[[Tsochantaridis et al, JMLR 2005| Tsochantaridis 
et 
al. 
(2005)]]
**Tsochantaridis 
et 
al. 
(2005): 

cutting 
planes 
(based 
on
 dual)

 
**Taskar
 et 
al. 
(2005):

 dual 
extragradient

 
**Ratliff
 et
 al.(2006):

 (stochastic) 
subgradient
 descent

 
**Crammer
 et 
al. 
(2006):

 online
“passive‐aggressive”
 algorithms
 

Latest revision as of 16:38, 2 November 2011

The Method and When to Use it

Structured (or Structural) Support Vector Machines (SSVM), as the name states, is a machine learning model that generalizes the Support Vector Machine (SVM) classifier, allowing training a classifier for structured output.

In general, SSVMs perform supervised learning by approximating a mapping

where is a set of labeled training examples and is a complex structured object, like trees, sequences, or sets, instead of simple univariate predictions (as in the SVM case).

Thus, training a SSVM classifier consists of showing pairs of correct sample and output label pairs, that are used for training, allowing to predict for new sample instances the corresponding output label

In NLP one can fing a great variety of problems that rely on complex outputs, such as parsing and Markov Models for part-of-speech tagging.

Training

While on training, for a set of samples and labels , the SSVM minimizes the risk function:


where is an arbitrary function, which measures the distance between to labels and is a function on samples and labels, which extracts feature vectors. These two function are to be defined according to the problem at hand


Since the equation above is non-differentiable, one can reformulate it introducing slack variables, , representing the value of the maximum. Using this approach the SSVM comes as:

      
      

Testing

Given a sample, and a mapping one can obtain the correspondent lable. The mapping is defined as


where is the vector during the training phase.

The inference problem of solving for this maximizer over the label space is dependent on the structure of the function that varies according to the problem.

Related Papers