Paper:Collins, ACL 2002

From Cohen Courses
Jump to navigationJump to search

Citation

Ranking Algorithms for Named-Entity Extraction: Boosting and the Voted Perceptron, Collins, ACL 2002

Online Version

Here is the online version of the paper.

Summary

In this paper the author describes two algorithms, wich rerank the top N hypotheses from a maximum entropy tagger, and apply them to the task of recovering named-entity boundaries. Since the task at hand can be framed as a tagging task - to tag each word as being either the start of an entity (S), a continuation of an entity (C), or not to be part of ab entity at all (N), the author uses state of the art maximum entropy tagger similar to the one described in Ratnaparkhi EMNLP 1996 as a baseline model. The same Max-ent tagger is used to generate the N=20 most probable segmentations for each input sentence, along with their probabilities. Author's aim is to come up with reranking strategies for the test data candidates in order to improve the precision and recall.

The author considers various global features for each candidate tagged sequence. Most of these features are anchored on entity boundaries in the candidate segmentation. Each candidate tagged sequence , proposed by the Max-ent tagger, is represented by the log probability from the tagger, as well as the values of the global features for ... , where are the no. of global features. The two algorithms described in the next section blend these two sources of information (global features and log probability) to improve upon a strategy which just takes the candidate from the tagger with the highest score for .

The author has shown that existing reranking methods - boosting and voted perceptron, are useful for a new domain, named-entity tagging. Another contribution is the suggestion of various global features of the candidate segmentations which give improvements on this task.

Brief description of the methods

Pairs of the form , where each is a sentence and each is the correct tag sequence for that sentence, serve as the training data. denotes the set of candidates (top N outputs from the Max-ent tagger) for each , where is the th candidate of th sentence. is the probability that the base model assigns to . Hence =log. The set of global features are represented as for ... . The parameters of the model are a vector of parameters, . The ranking function is given as

The ranking function can also be written as if is defined as .

Boosting for ranking

The boosting algorithm in the paper is the one described in Collins 2000, where the method can be considered to be a greedy algorithm for finding the parameters that minimize the loss function

As an initial step, is set to be

and all other parameters for ... are set to be zero. The algorithm then proceeds for N iterations. At each iteration, a single feature is chosen, and its weight is updated. Suppose the current parameter values are , and a single feature is chosen, its weight being updated through an increment , i.e., . Then the new loss, after this parameter update, will be

where . The boosting algorithm chooses the feature/update pair which is optimal in terms of minimizing the loss function, i.e.,

and then makes the update .

Voted Perceptron for ranking

The following steps define the perceptron training algorithm for ranking problems.

Define:

Input: Examples with feature vectors

Initialization: Set parameters

For


If  Then 
         Else 

Output: Parameter vectors for

The algorithm maintains a parameter vector , which is initially set to be all zeros. The algorithm then makes a pass over the training set, at each training example storing a parameter vector for . The parameter vector is only modified when a mistake is made on an example. The update here involves adding the difference of the offending examples' representations. In the voted perceptron all parameter vectors for are stored. Each of these parameter settings "vote" for a candidate, which is the highest ranking candidate of each parameter setting given by , where . The candidate with the highest no. of votes is returned as the most likely candidate.

Experimental Result

Following table shows the Precision (P), Recall (R) and F-measure (F) for the three tagging methods. Figures in parantheses are the relative improvements in error rate over the Max-ent model. All figures are percentages.

Collins results table.png

Related papers

Collins 2000 used boosting for to rerank the output of an existing probabilistic parser and compared with Markov Random fields.