Difference between revisions of "Koehn et al, ACL 2003"
Line 45: | Line 45: | ||
The second method explored act as a filter to the previous set of alignments, restricting possible phrases to syntactically correct ones. | The second method explored act as a filter to the previous set of alignments, restricting possible phrases to syntactically correct ones. | ||
− | Finally, the last method | + | Finally, the last method takes the [[RelatedPaper::Marcus and Wong, EMNLP 2002]] approach, learning phrase-level alignments directly from the parallell corpora. |
− | |||
− | [[RelatedPaper:: | ||
== Experimental Results == | == Experimental Results == | ||
The authors used the [[UsesDataset::Hansards]] | The authors used the [[UsesDataset::Hansards]] | ||
[[File:Lacostejulienresults.png|300px]] | [[File:Lacostejulienresults.png|300px]] |
Revision as of 18:46, 28 November 2011
Being edited by Rui Correia
Contents
Citation
Philipp Koehn, Franz Josef Och, and Daniel Marcu. 2003. Statistical phrase-based translation. In Proceedings of HLT-NAACL 2003, pages 127–133. [1]
Summary
In this paper the authors propose a new framework that aims at explaining and understanding why phrase-based models in Machine Translation outperform word-based models.
Within this framework (phrase-based translation model and decoding algorithm) the authors carry experiments that explore three different methods for learning phrase translation (based on word alignments, on syntactic information and "pure" phrase alignments). Additionally the authors also explore phrase length, lexical weighting, and the impact of different language pairs in the overall BLEU score.
The results confirm the already proved hypotheses that phrase translation achieve better performance than word-based methods, adding that three-word phrase are sufficient to outperform the traditional methods. Moreover, the authors conclude that lexical weighting of phrase translation boost results, and that syntactic considerations, on the other hand, hinder the results.
Evaluation Framework
The phrase translation model used in the proposed framework is based on the noisy channel model. The best English output sentence given a foreign input sentence is given by:
where:
- is the translation model (see below);
- is a trigram language model;
- and, is a factor that calibrates the output length (\omega > 1, biasing longer output).
The translation model can be decomposed into:
where:
- is a sequence of segmented from the input sentence ;
- is a probability distribution that models the phrase translation;
- and, is a relative distortion probability distribution between the start position of the foreign phrase that was translated into the th English phrase () and the end position of the foreign phrase translated into the th English phrase ().
The decoder that was adopted in the framework employs a Beam Search algorithm.
Methods for Learning Phrase Translation
In this work the authors compare three methods to build phrase translation probability tables. The first one builds the phrase alignments using word alignment information, i.e., all the phrase pairs that are considered must be consistent with the word alignments.
The second method explored act as a filter to the previous set of alignments, restricting possible phrases to syntactically correct ones.
Finally, the last method takes the Marcus and Wong, EMNLP 2002 approach, learning phrase-level alignments directly from the parallell corpora.
Experimental Results
The authors used the Hansards