Difference between revisions of "Paper:Collins, ACL 2002"

From Cohen Courses
Jump to navigationJump to search
Line 36: Line 36:
 
and then makes the update <math>w_{k^*} = w_{k^*} + \delta^*</math>.
 
and then makes the update <math>w_{k^*} = w_{k^*} + \delta^*</math>.
 
=== [[Voted Perceptron]] for ranking ===
 
=== [[Voted Perceptron]] for ranking ===
 +
The following steps define the perceptron training algorithm for ranking problems.
  
 +
'''Define:''' <math>F(x,w)=w.h(x)</math>
 +
 +
'''Input:''' Examples <math>x_{ij}</math> with feature vectors <math>h(x_{ij})</math>
 +
 +
'''Initialization:''' Set parameters <math>w^0=0</math>
 +
 +
'''For''' <math>i = 1 \ldots n</math>
 +
 +
<math>j = argmax_{j=1...n_{i}} F(x_{ij},w^{i-1})</math>
 +
'''If''' <math>(j=1)</math> '''Then''' <math>w^i = w^{i-1}</math>
 +
          '''Else''' <math>w^i = w^{i-1} + h(x_{i1}) - h(x_{ij})</math>
 +
 +
'''Output:''' Parameter vectors <math>w^i</math> for <math>i = 1 \ldots n</math>
 
== Experimental Result ==
 
== Experimental Result ==
  
 
== Related papers ==
 
== Related papers ==

Revision as of 14:32, 26 September 2011

Citation

Ranking Algorithms for Named-Entity Extraction: Boosting and the Voted Perceptron, Collins, ACL 2002

Online Version

Here is the online version of the paper.

Summary

In this paper the author describes two algorithms, wich rerank the top N hypotheses from a maximum entropy tagger, and apply them to the task of recovering named-entity boundaries. Since the task at hand can be framed as a tagging task - to tag each word as being either the start of an entity (S), a continuation of an entity (C), or not to be part of ab entity at all (N), the author uses state of the art maximum entropy tagger similar to the one described in Ratnaparkhi EMNLP 1996 as a baseline model. The same Max-ent tagger is used to generate the N=20 most probable segmentations for each input sentence, along with their probabilities. Author's aim is to come up with reranking strategies for the test data candidates in order to improve the precision and recall.

The author considers various global features for each candidate tagged sequence. Most of these features are anchored on entity boundaries in the candidate segmentation. Each candidate tagged sequence , proposed by the Max-ent tagger, is represented by the log probability from the tagger, as well as the values of the global features for ... , where are the no. of global features. The two algorithms described in the next section blend these two sources of information (global features and log probability) to improve upon a strategy which just takes the candidate from the tagger with the highest score for .

The author has shown that existing reranking methods - boosting and voted perceptron, are useful for a new domain, named-entity tagging. Another contribution is the suggestion of various global features of the candidate segmentations which give improvements on this task.

Brief description of the methods

Pairs of the form , where each is a sentence and each is the correct tag sequence for that sentence, serve as the training data. denotes the set of candidates (top N outputs from the Max-ent tagger) for each , where is the th candidate of th sentence. is the probability that the base model assigns to . Hence =log. The set of global features are represented as for ... . The parameters of the model are a vector of parameters, . The ranking function is given as

The ranking function can also be written as if is defined as .

Boosting for ranking

The boosting algorithm in the paper is the one described in Collins 2000, where the method can be considered to be a greedy algorithm for finding the parameters that minimize the loss function

As an initial step, is set to be

and all other parameters for ... are set to be zero. The algorithm then proceeds for N iterations. At each iteration, a single feature is chosen, and its weight is updated. Suppose the current parameter values are , and a single feature is chosen, its weight being updated through an increment , i.e., . Then the new loss, after this parameter update, will be

where . The boosting algorithm chooses the feature/update pair which is optimal in terms of minimizing the loss function, i.e.,

and then makes the update .

Voted Perceptron for ranking

The following steps define the perceptron training algorithm for ranking problems.

Define:

Input: Examples with feature vectors

Initialization: Set parameters

For


If  Then 
         Else 

Output: Parameter vectors for

Experimental Result

Related papers