Difference between revisions of "Accurate Unlexicalized Parsing"

From Cohen Courses
Jump to navigationJump to search
Line 174: Line 174:
 
== Related Work ==
 
== Related Work ==
  
Coming soon!
+
Many strides have been made in unlexicalized parsing since this paper (many of which are by the authors of this paper).
 +
 
 +
* [[RelatedPaper::Improved Inference for Unlexicalized Parsing, Slav Petrov and Dan Klein, HLT-NAACL, 2007]] - Improves upon (clearly) the inference of unlexicalized parsing
 +
* [[RelatedPaper::Learning Accurate, Compact, and Interpretable Tree Annotation, S. Petrov, L. Barrett, R. Thibaux, D. Klein, ACL 2006]] - In this paper, they learn tree annotations to actually produce better than lexicalized results

Revision as of 19:56, 1 November 2011

This paper is a work in progress by Francis Keith

Citation

"Accurate Unlexicalized Parsing", D. Klein and C. D. Manning, ACL 2003

Online Version

An online version of this paper is available here [1]

Summary

One of the concepts in PCFG parsing that has become fairly standard is the concept of lexicalizing the grammar; that is, annotating each node with a head word. The authors of the paper focus on using unlexicalized grammars, and attempt to exploit structural context in an effort to produce improved results that can potentially be combined with lexicalization to improve the state-of-the-art. They apply some linguistic knowledge to modify the structure of the grammars as well.

Lexicalized vs Unlexicalized

Some of the methods they use seem very similar to what goes on in a lexicalized PCFG. An important differentiation between the two comes in about content words vs function words. The argument is made that linguists will often differentiate between annotating nodes that have a different functional head, as opposed to a different content head. That is, the authors are attempting to leverage the linguistic structure of a phrase, not leverage past information obtained from training data. So is structurally different from from a linguistic standpoint, but to annotate using content words, like vs is what the authors are seeking to avoid.

Method

They do an iterative set of non-lexicalization improvements on the grammars and report the improving F1 scores. They use the standard CKY parsing algorithm for running their experiments. The grammar probabilities are given as unsmoothed maximum likelihood probabilities.

Markovization

The first step that they take is to markovize the rules. In markovization, we examine the vertical and horizontal ancestors of the current node. It is logical to take vertical, and horizontal ancestors. In the original TreeBank, , because only the current node is considered vertically (i.e., no parent node history stored), and the node is based on all the nodes in the rewrite rule that came before it.

In their experiments, they found that was optimal, because, while it did not perform the best of the tests they did, it had significantly fewer symbols which would allow for them to do further rule splitting and still keep the grammar size managable. It is essentially the best compromise between improved performance from markovization, and grammar complexity.

Tag Splitting

The remainder of the methods involve splitting the grammar up and adding additional annotations.

They added various annotations for:

  • UNARY-Internal: Any nonterminal with only one child
  • UNARY-External: Any nonterminal that is an only child (discarded)
  • UNARY-DT: Any determiner that is an only child (subset of Unary-External that provided a benefit)
  • UNARY-RB: Any adverb that is an only child (subset of Unary-External that provided a benefit)
  • TAG-PA: They included an annotation for each node's parent node
  • SPLIT-IN: They split the IN annotation up based on linguistic knowledge
  • SPLIT-CC: They split conjunctions into but/and and others (because those have different linguistic properties)
  • SPLIT-AUX: They split auxiliary verbs into "be" and "have"
  • TMP-NP: Keeps the temporal tag on noun phrases
  • GAPPED-S: Marks S nodes that have an empty subject
  • POSS-NP: Marks possessive NPs
  • SPLIT-VP: Splits the VP annotation to mark with the head word. They then merge all the finite forms into VBF, to avoid doing lexicalization
    • Comment: This seems almost like lexicalization...
  • BASE-NP: Annotate non-recursive NPs
  • DOMINATES-V: Mark nodes that dominate the verb
  • RIGHT-REC-NP: Mark all NPs that contain an NP as a right most descendant
    • This is an attempt to model depth

Note that all of these annotations are not caused by lexicalization (with the possible exception of SPLIT-VP), but more based on linguistic understanding.

Experiments

They do the parsing evaluation using the Penn Treebank. They use sections 2-21 for training and 22 for development. Section 23 is used as the test set. They report the scores for each additional component rule added on the development set, as well as the final score on the test set.

In this case, the baseline results are obtained with a markovization of . The changes in are given with the annotation applied both cumulatively, and also in isolation.

Results on the development set:

Annotation Size (total) (individual)
Baseline 7619 77.77 N/A N/A
UNARY-INTERNAL 8065 78.32 0.55 0.55
UNARY-DT 8066 78.48 0.71 0.17
UNARY-RB 8069 78.86 1.09 0.43
TAG-PA 8520 80.62 2.85 2.52
SPLIT-IN 8541 81.19 3.42 2.12
SPLIT-AUX 9034 81.66 3.89 0.57
SPLIT-CC 9190 81.69 3.92 0.12
SPLIT-% 9255 81.81 4.04 0.14
TMP-NP 9594 82.25 4.48 1.07
GAPPED-S 9741 82.28 4.51 0.17
POSS-NP 9820 83.06 5.29 0.28
SPLIT-VP 10499 85.72 7.95 1.36
BASE-NP 11660 86.04 8.27 0.73
DOMINATES-V 14097 86.91 9.14 1.42
RIGHT-REC-NP 15276 87.04 9.27 1.94

They get a final test set of 68.32%, which is close to that of early lexicalized parsers, but a bit lower than more recent lexicalized parsers.

Conclusion

Using only some grammar structure manipulations, they were able to get vast improvement over the baseline. It's possible there could be some intelligent way to leverage the linguistic structure that they use and combine it with lexical features.

Related Work

Many strides have been made in unlexicalized parsing since this paper (many of which are by the authors of this paper).