Difference between revisions of "Dyer et al, ACL 2011"
(→Model) |
|||
(2 intermediate revisions by the same user not shown) | |||
Line 34: | Line 34: | ||
<math> | <math> | ||
− | Z_{\theta} (s,n)= \sum_{t' \in \Omega ^n} \ | + | Z_{\theta} (s,n)= \sum_{t' \in \Omega ^n} \sum_{a'} \exp \theta ^\top H(t',a',s,n) |
</math> | </math> | ||
== Features == | == Features == | ||
− | + | The authors group the result of the feature engineering task in 5 feature categories. | |
+ | |||
+ | * The first group described is comprised of <b>Word Association Features</b>, which are common to all lexical translation models, containing features such as fine-grained boolean indicators, orthographic similarities or class pair indicator features. | ||
+ | |||
+ | *The second group, <b>Positional Features</b>, is responsible to represent closeness to the alignment matrix diagonal, and when conjoined with the class indicators from the previous group, are able to represent the typical location of certain classes on the sentences. | ||
+ | |||
+ | *<b>Source Features</b> represent the need to translate certain words more than others (eg. functional elements of the source language). | ||
+ | |||
+ | *The fourth set of features is composed by <b>Source Path Features</b> that are able to represent typical sentence structure and ordering, which when combined with word classes, for instance, can represent the common position of an adjective with respect to a noun in a given language. | ||
+ | |||
+ | *The last group, <b>Target String Features</b>, is responsible to capture the phenomenon of multiple values in the predicted target string, and is fired when, for example, a word translates as itself in a given position, but then is translated again as something else in the previous or next positions. | ||
== Experimental Results == | == Experimental Results == | ||
− | [[UsesDataset:: | + | To assess their model the authors used corpora collected from different sources. For the Chinese-English language pair, they used a tourism domain corpus from [http://www.lrec-conf.org/proceedings/lrec2002/pdf/305.pdf]. The results in terms of BLEU, METEOR and TER are shown in the table below, where the new model is compared with the IBM Model 4, and with a merging between these two strategies. |
+ | |||
+ | [[File:Dyerchineseenglish.png|300px]] | ||
+ | |||
+ | |||
+ | For the Czech-English and Urdu-English language pairs, the authors used the [[UsesDataset::NIST MT]] dataset. The results for these two language pairs can be found in the following tables: | ||
− | [[File: | + | [[File:Dyerczechenglish.png|300px]] [[File:Dyerurduenglish.png|300px]] |
Latest revision as of 14:12, 29 November 2011
Being edited by Rui Correia
Citation
C. Dyer, J. Clark, A. Lavie, and N. A. Smith. 2011. Unsupervised Word Alignment with Arbitrary Features. In Proceedings of HLT-ACL 2011, Volume 1, pp 409–419.
Summary
In this paper the authors address the Word Alignments problem in an unsupervised fashion, filling the gap of having to manually develop a gold standard that is difficult and expensive to create and dependent on the task at hand, specially in languages with resource scarcity problems. The model that is introduced is discriminatively trained, globally normalized, being a variant of the IBM Model 1 that allows the incorporation of non-independent features.
The main focus of the paper goes to the new model propose and to the features that were considered to generate the word alignments. The authors show results for several language pairs, comparing their approach with the IBM Model 4 with respect to BLEU, METEOR and TER scores. Additionally, the authors look at how the different language pairs use the features that were designed in different ways, analyzing how these preferences are representative of each language.
Model
The conditional model proposed assigns probabilities to a target sentence with length , given a source language sentence , with length . Using the chain rule, the authors factor in a translation model and a length model , i.e.,
Regarding the translation model, the authors make the assumption that each word of the target language sentence is the translation of a single word in the source language or a special null token, introducing a latent variable , i.e.,
It is at this point that the model diverges from the Brown et al. approach. Instead of using the chain rule, the author propose the application of a log-linear model with parameters and feature vector function that maps each tuple into to model directly:
where, is the partition function given by
Features
The authors group the result of the feature engineering task in 5 feature categories.
- The first group described is comprised of Word Association Features, which are common to all lexical translation models, containing features such as fine-grained boolean indicators, orthographic similarities or class pair indicator features.
- The second group, Positional Features, is responsible to represent closeness to the alignment matrix diagonal, and when conjoined with the class indicators from the previous group, are able to represent the typical location of certain classes on the sentences.
- Source Features represent the need to translate certain words more than others (eg. functional elements of the source language).
- The fourth set of features is composed by Source Path Features that are able to represent typical sentence structure and ordering, which when combined with word classes, for instance, can represent the common position of an adjective with respect to a noun in a given language.
- The last group, Target String Features, is responsible to capture the phenomenon of multiple values in the predicted target string, and is fired when, for example, a word translates as itself in a given position, but then is translated again as something else in the previous or next positions.
Experimental Results
To assess their model the authors used corpora collected from different sources. For the Chinese-English language pair, they used a tourism domain corpus from [1]. The results in terms of BLEU, METEOR and TER are shown in the table below, where the new model is compared with the IBM Model 4, and with a merging between these two strategies.
For the Czech-English and Urdu-English language pairs, the authors used the NIST MT dataset. The results for these two language pairs can be found in the following tables: