McDonald et al, ACL 2005: Non-Projective Dependency Parsing Using Spanning Tree Algorithms

From Cohen Courses
Revision as of 17:28, 22 October 2011 by Yunwang (talk | contribs)
Jump to navigationJump to search

Citation

R. McDonald, F. Pereira, K. Ribarov, J. Hajič. Non-projective dependency parsing using spanning tree algorithms, Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing, pp. 523-530, Vancouver, October 2005.

Online Version

PDF version

Summary

This paper addresses the problem of non-projective dependency prasing.

Given a sentence , we can construct a directed graph . The vertex set contains one vertex for each word in the sentence, plus a dummy vertex for the "root". The edge set contains all directed edges of the form , where , , and . Each edge has a score of the form , where is a feature vector depending on the words and , and is a weight factor.

A (non-projective) dependency parse tree is a subgraph of which covers all the vertices in , and in which each vertex has exactly one predecessor (except for the root vertex which has no predecessor). The score of a dependency tree is factored as the sum of the scores of its edges:

The decoding problem is to find the max-scoring dependency parse tree given a sentence , assuming that the weight vector is known. The learning problem is to find an optimal weight vector , such that the sum of the scores of the dependency parse trees in a training corpus is maximized.

The Decoding Problem

The Learning Problem

Experiments