Difference between revisions of "Accurate Unlexicalized Parsing"
Line 55: | Line 55: | ||
== Experiments == | == Experiments == | ||
− | They do the parsing evaluation using the [[UsesDataset::Penn Treebank]]. They use sections 2-21 for training and 22 for development. Section 23 is | + | They do the parsing evaluation using the [[UsesDataset::Penn Treebank]]. They use sections 2-21 for training and 22 for development. Section 23 is used as the test set. They report the <math>F_{1}</math> scores for each additional component rule added on the development set, as well as the final score on the test set. |
+ | |||
+ | In this case, the baseline results are obtained with a markovization of <math>v\leq2,h\leq2</math>. The changes in <math>F_{1}</math> are given with the annotation applied both cumulatively, and also in isolation. | ||
+ | |||
+ | Results on the development set: | ||
+ | |||
+ | {| class="wikitable" border="1" | ||
+ | |- | ||
+ | ! Annotation | ||
+ | ! Size | ||
+ | ! <math>F_{1}</math> | ||
+ | ! <math>\Delta F_{1}</math> (total) | ||
+ | ! <math>\Delta F_{1}</math> (individual) | ||
+ | |- | ||
+ | ! Baseline | ||
+ | | 7619 | ||
+ | | 77.77 | ||
+ | | N/A | ||
+ | | N/A | ||
+ | |- | ||
+ | ! UNARY-INTERNAL | ||
+ | | 8065 | ||
+ | | 78.32 | ||
+ | | 0.55 | ||
+ | | 0.55 | ||
+ | |- | ||
+ | ! UNARY-DT | ||
+ | | 8066 | ||
+ | | 78.48 | ||
+ | | 0.71 | ||
+ | | 0.17 | ||
+ | |- | ||
+ | ! UNARY-RB | ||
+ | | 8069 | ||
+ | | 78.86 | ||
+ | | 1.09 | ||
+ | | 0.43 | ||
+ | |- | ||
+ | ! TAG-PA | ||
+ | | 8520 | ||
+ | | 80.62 | ||
+ | | 2.85 | ||
+ | | 2.52 | ||
+ | |- | ||
+ | ! SPLIT-IN | ||
+ | | 8541 | ||
+ | | 81.19 | ||
+ | | 3.42 | ||
+ | | 2.12 | ||
+ | |- | ||
+ | ! SPLIT-AUX | ||
+ | | 9034 | ||
+ | | 81.66 | ||
+ | | 3.89 | ||
+ | | 0.57 | ||
+ | |- | ||
+ | ! SPLIT-CC | ||
+ | | 9190 | ||
+ | | 81.69 | ||
+ | | 3.92 | ||
+ | | 0.12 | ||
+ | |- | ||
+ | ! SPLIT-% | ||
+ | | 9255 | ||
+ | | 81.81 | ||
+ | | 4.04 | ||
+ | | 0.14 | ||
+ | |- | ||
+ | ! TMP-NP | ||
+ | | 9594 | ||
+ | | 82.25 | ||
+ | | 4.48 | ||
+ | | 1.07 | ||
+ | |- | ||
+ | ! GAPPED-S | ||
+ | | 9741 | ||
+ | | 82.28 | ||
+ | | 4.51 | ||
+ | | 0.17 | ||
+ | |- | ||
+ | ! POSS-NP | ||
+ | | 9820 | ||
+ | | 83.06 | ||
+ | | 5.29 | ||
+ | | 0.28 | ||
+ | |- | ||
+ | ! SPLIT-VP | ||
+ | | 10499 | ||
+ | | 85.72 | ||
+ | | 7.95 | ||
+ | | 1.36 | ||
+ | |- | ||
+ | ! BASE-NP | ||
+ | | 11660 | ||
+ | | 86.04 | ||
+ | | 8.27 | ||
+ | | 0.73 | ||
+ | |- | ||
+ | ! DOMINATES-V | ||
+ | | 14097 | ||
+ | | 86.91 | ||
+ | | 9.14 | ||
+ | | 1.42 | ||
+ | |- | ||
+ | ! RIGHT-REC-NP | ||
+ | | 15276 | ||
+ | | 87.04 | ||
+ | | 9.27 | ||
+ | | 1.94 | ||
+ | |} |
Revision as of 19:26, 1 November 2011
This paper is a work in progress by Francis Keith
Contents
Citation
"Accurate Unlexicalized Parsing", D. Klein and C. D. Manning, ACL 2003
Online Version
An online version of this paper is available here [1]
Summary
One of the concepts in PCFG parsing that has become fairly standard is the concept of lexicalizing the grammar; that is, annotating each node with a head word. The authors of the paper focus on using unlexicalized grammars, and attempt to exploit structural context in an effort to produce improved results that can potentially be combined with lexicalization to improve the state-of-the-art. They apply some linguistic knowledge to modify the structure of the grammars as well.
Lexicalized vs Unlexicalized
Some of the methods they use seem very similar to what goes on in a lexicalized PCFG. An important differentiation between the two comes in about content words vs function words. The argument is made that linguists will often differentiate between annotating nodes that have a different functional head, as opposed to a different content head. That is, the authors are attempting to leverage the linguistic structure of a phrase, not leverage past information obtained from training data. So is structurally different from from a linguistic standpoint, but to annotate using content words, like vs is what the authors are seeking to avoid.
Method
They do an iterative set of non-lexicalization improvements on the grammars and report the improving F1 scores. They use the standard CKY parsing algorithm for running their experiments. The grammar probabilities are given as unsmoothed maximum likelihood probabilities.
Markovization
The first step that they take is to markovize the rules. In markovization, we examine the vertical and horizontal ancestors of the current node. It is logical to take vertical, and horizontal ancestors. In the original TreeBank, , because only the current node is considered vertically (i.e., no parent node history stored), and the node is based on all the nodes in the rewrite rule that came before it.
In their experiments, they found that was optimal, because, while it did not perform the best of the tests they did, it had significantly fewer symbols which would allow for them to do further rule splitting and still keep the grammar size managable. It is essentially the best compromise between improved performance from markovization, and grammar complexity.
Tag Splitting
The remainder of the methods involve splitting the grammar up and adding additional annotations.
They added various annotations for:
- UNARY-Internal: Any nonterminal with only one child
- UNARY-External: Any nonterminal that is an only child (discarded)
- UNARY-DT: Any determiner that is an only child (subset of Unary-External that provided a benefit)
- UNARY-RB: Any adverb that is an only child (subset of Unary-External that provided a benefit)
- TAG-PA: They included an annotation for each node's parent node
- SPLIT-IN: They split the IN annotation up based on linguistic knowledge
- SPLIT-CC: They split conjunctions into but/and and others (because those have different linguistic properties)
- SPLIT-AUX: They split auxiliary verbs into "be" and "have"
- TMP-NP: Keeps the temporal tag on noun phrases
- GAPPED-S: Marks S nodes that have an empty subject
- POSS-NP: Marks possessive NPs
- SPLIT-VP: Splits the VP annotation to mark with the head word. They then merge all the finite forms into VBF, to avoid doing lexicalization
- Comment: This seems almost like lexicalization...
- BASE-NP: Annotate non-recursive NPs
- DOMINATES-V: Mark nodes that dominate the verb
- RIGHT-REC-NP: Mark all NPs that contain an NP as a right most descendant
- This is an attempt to model depth
Note that all of these annotations are not caused by lexicalization (with the possible exception of SPLIT-VP), but more based on linguistic understanding.
Experiments
They do the parsing evaluation using the Penn Treebank. They use sections 2-21 for training and 22 for development. Section 23 is used as the test set. They report the scores for each additional component rule added on the development set, as well as the final score on the test set.
In this case, the baseline results are obtained with a markovization of . The changes in are given with the annotation applied both cumulatively, and also in isolation.
Results on the development set:
Annotation | Size | (total) | (individual) | |
---|---|---|---|---|
Baseline | 7619 | 77.77 | N/A | N/A |
UNARY-INTERNAL | 8065 | 78.32 | 0.55 | 0.55 |
UNARY-DT | 8066 | 78.48 | 0.71 | 0.17 |
UNARY-RB | 8069 | 78.86 | 1.09 | 0.43 |
TAG-PA | 8520 | 80.62 | 2.85 | 2.52 |
SPLIT-IN | 8541 | 81.19 | 3.42 | 2.12 |
SPLIT-AUX | 9034 | 81.66 | 3.89 | 0.57 |
SPLIT-CC | 9190 | 81.69 | 3.92 | 0.12 |
SPLIT-% | 9255 | 81.81 | 4.04 | 0.14 |
TMP-NP | 9594 | 82.25 | 4.48 | 1.07 |
GAPPED-S | 9741 | 82.28 | 4.51 | 0.17 |
POSS-NP | 9820 | 83.06 | 5.29 | 0.28 |
SPLIT-VP | 10499 | 85.72 | 7.95 | 1.36 |
BASE-NP | 11660 | 86.04 | 8.27 | 0.73 |
DOMINATES-V | 14097 | 86.91 | 9.14 | 1.42 |
RIGHT-REC-NP | 15276 | 87.04 | 9.27 | 1.94 |