Difference between revisions of "Models of metaphor in NLP"

From Cohen Courses
Jump to navigationJump to search
Line 58: Line 58:
  
 
=== Krishnakumaran & Zhu (2007) ===
 
=== Krishnakumaran & Zhu (2007) ===
Both Birke and Sarkar (2006) and Gedigan et
+
* Deal with not only verbs, but also '''nouns and adjectives'''.
al. (2006) focus only on metaphors expressed by
+
* Use hyponymy relation in WordNet and word bigram counts to predict metaphors at sentence level.
a verb. As opposed to that the approach of Krishnakumaran
+
* Approach
and Zhu (2007) deals with verbs,
+
** '''Noun-Noun''': Given an '''IS-A''' metaphor, verify if the two nouns involved are in hyponymy relation in WordNet, if not, tag as a metaphor
nouns and adjectives as parts of speech. They
+
** '''Verb-Noun and Adjective-Noun''': Calculate bigram probabilities of verb-noun and adj-noun pairs, if the pair has low frequency, tag as a metaphor
use hyponymy relation in WordNet and word bigram
+
* Fail to capture conventionalized metaphors (loose information compared of parsed text)
counts to predict metaphors at a sentence
 
level. Given an IS-A metaphor (e.g. The world
 
is a stage3) they verify if the two nouns involved
 
are in hyponymy relation in WordNet, and if
 
they are not then this sentence is tagged as containing
 
a metaphor. Along with this they consider
 
expressions containing a verb or an adjective
 
used metaphorically (e.g. He planted good
 
ideas in their minds or He has a fertile imagination).
 
Hereby they calculate bigram probabilities
 
of verb-noun and adjective-noun pairs (including
 
the hyponyms/hypernyms of the noun in
 
question). If the combination is not observed in
 
the data with sufficient frequency, the system tags
 
the sentence containing it as metaphorical. This
 
idea is a modification of the selectional preference
 
view of Wilks. However, by using bigram
 
counts over verb-noun pairs Krishnakumaran and
 
Zhu (2007) loose a great deal of information compared
 
to a system extracting verb-object relations
 
from parsed text. The authors evaluated their system
 
on a set of example sentences compiled from
 
the Master Metaphor List (Lakoff et al., 1991),
 
whereby highly conventionalized metaphors (they
 
call them dead metaphors) are taken to be negative
 
examples. Thus they do not deal with literal examples
 
as such: essentially, the distinction they are
 
making is between the senses included in Word-
 
Net, even if they are conventional metaphors, and
 
those not included in WordNet.
 
  
 
== Metaphor Interpretation ==
 
== Metaphor Interpretation ==

Revision as of 20:01, 8 October 2012

Citation

E. Shutova. 2010. Models of Metaphor in NLP. In Proceedings of ACL 2010, Uppsala, Sweden.

Online version

ACL anthology

Introduction

This is a review paper of modeling metaphors in NLP. The author devised it into two main tasks: "metaphor recognition" and "metaphor interpretation".

Metaphor Recognition

Met* System (Fass, 1991)

  • First attempt to identify and interpret metaphorical expression
  • Using selectional preference and hand-coded knowledge base
  • 3-Stage Approaches
  1. Detect selectional preference violation
  2. If find violations, tested for being a metonymic relation using hand-coded patterns
  3. If not metonymy, search the knowledge base for a relevant analogy in order to discriminate metaphorical relations
  • Problem
  1. Detects any kind of non-literalness in language (metaphors, metonymies and others), and not only metaphors
  2. Fail to detect high conventionality of metaphor

Goatly (1997)

  • Identify a set of linguistic cues indicate metaphor
    • metaphorically speaking, utterly, completely, so to speak and, surprisingly, literally.

Peters & Peters (2000)

  • Detect figurative language in the WordNet
  • Search for systematic polysemy, which allows to capture metonymic and metaphorical relations

CorMet System (Mason, 2004)

  • The first attempt for source-target domain mapping
  • A corpus-based approach to find systematic variations in domain-specific selectional preferences
  • Take Master Metaphor List (Lakoff et al., 1991) as baseline, and achieve an accuracy of 77% (judged by human)

TroFi System (Birke & Sarkar, 2006)

  • Sentence clustering approach for non-literal language recognition
  • Inspired by a similarity-based word sense disambiguation method
  • Approach
  1. Use a set of seed sentences (human annotated)
  2. Compute similarity between (1) the sentence containing the word to be disambiguated and (2) all of the seed sentences
  3. Select the sense corresponding to the annotation in the most similar seed sentences
  • F1-score = 0.538. But the task is not clearly defined.

Gedigan et al. (2006)

  • A survey-like work about literal and metaphorical uses of verbs
  • Method
  1. Extract the lexical items whose frames are related to MOTION and CURE from FrameNet (Fillmore et al., 2003).
  2. Search the PropBank Wall Street Journal corpus (Kingsbury and Palmer, 2002) for sentences containing such lexical items and annotated them with respect to metaphoricity (by hand).
  3. Used PropBank annotation as features to train the maximum entropy classifier
  • Accuracy = 95.12%
    • Very close to majority baseline (92.90%) => 92.00% of the verbs of MOTION and CURE in the Wall Street Journal corpus are used metaphorically.

Krishnakumaran & Zhu (2007)

  • Deal with not only verbs, but also nouns and adjectives.
  • Use hyponymy relation in WordNet and word bigram counts to predict metaphors at sentence level.
  • Approach
    • Noun-Noun: Given an IS-A metaphor, verify if the two nouns involved are in hyponymy relation in WordNet, if not, tag as a metaphor
    • Verb-Noun and Adjective-Noun: Calculate bigram probabilities of verb-noun and adj-noun pairs, if the pair has low frequency, tag as a metaphor
  • Fail to capture conventionalized metaphors (loose information compared of parsed text)

Metaphor Interpretation

MIDAS System (Martin, 1990)

KARMA System (Narayanan, 1997), ATT-Meta (Barnden and Lee, 2002)

Veale and Hao (2008)

Shutova (2010)

Metaphor Resources

Metaphor Annotation in Corpora

Metaphor & Polysemy

Metaphor Identification

Pragglejaz Procedure

Source - Target Domain Vocabulary

Annotating Source and target Domains

Related papers

The widely cited Pang et al EMNLP 2002 paper was influenced by this paper - but considers supervised learning techniques. The choice of movie reviews as the domain was suggested by the (relatively) poor performance of Turney's method on movies.

An interesting follow-up paper is Turney and Littman, TOIS 2003 which focuses on evaluation of the technique of using PMI for predicting the semantic orientation of words.