Difference between revisions of "Hoffmann et al., ACL 2010"

From Cohen Courses
Jump to navigationJump to search
Line 26: Line 26:
  
 
== Brief description of the method ==
 
== Brief description of the method ==
 +
=== Schema Classifier ===
 +
The system uses a liner, multi-class classifier with 6 kinds of features:
 +
* Words in the article title;
 +
* Words in the first sentence;
 +
* Words in the first sentence which are direct objects to the verb 'to be' (''what is that?'')
 +
* Article section headers;
 +
* Wikipedia categories;
 +
* Ancestor categories.
 +
[Voted_Perceptron|Voted Perceptron] is used for training the liner classifier.
 +
=== Extractor ===
  
 
== Experimental Result ==
 
== Experimental Result ==
  
 
== Related papers ==
 
== Related papers ==

Revision as of 13:52, 30 September 2011

Citation

Raphael Hoffmann, Congle Zhang, and Daniel S. Weld. 2010. Learning 5000 relational extractors. In ACL '10 (Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics)

Online version

ACM Digital Library

ACL Anthology

PPT slides

Summary

This is a paper introducing LUCHS, a self-supervised relations-specific IE system capable of learning more than 5000 relations with an average F1 score of 61%. The system applies dynamic lexicon features learning is applied as a semi-supervise learning solution cope with sparse training data.

System Architecture

The following figure summarizes the architecture of LUCHS.

LUCHSArchitecture.png

  • A Schema Classifier is trained by Wikipedia pages containing infoboxes, which can be used to decide which schema should be applied to an article without infobox;
  • Training Data is generated heuristically by the Matcher (e.g. the Wikipedia article "Jerry Seinfeld" contains sentence "Seinfeld was born in Brooklyn, New York." and at the infobox within the same page, contains relation pair "birth_place = Brooklyn"; the Matcher heuristically generate training data for extractors of different relations (but in this paper the authors didn't describe any details about this part);
  • A language model is trained by CRF Learner, using the training data generated by Matcher, and used as a Extractor to extract structured information from free text;
  • As the major contribution of this paper, a Lexicon Learner is trained by HTML lists crawled from Internet, and contribute lexicons feature to CRF Learner. This step enables the system working with "sparse relations", or say extract structured information in a semi-supervised way.

Brief description of the method

Schema Classifier

The system uses a liner, multi-class classifier with 6 kinds of features:

  • Words in the article title;
  • Words in the first sentence;
  • Words in the first sentence which are direct objects to the verb 'to be' (what is that?)
  • Article section headers;
  • Wikipedia categories;
  • Ancestor categories.

[Voted_Perceptron|Voted Perceptron] is used for training the liner classifier.

Extractor

Experimental Result

Related papers