Difference between revisions of "Koehn et al, ACL 2003"

From Cohen Courses
Jump to navigationJump to search
Line 48: Line 48:
  
 
== Experimental Results ==
 
== Experimental Results ==
The authors used the [[UsesDataset::Hansards]]  
+
The authors used the [[UsesDataset::EUROPARL]] for the pair German-English.
 +
 
 
[[File:Lacostejulienresults.png|300px]]
 
[[File:Lacostejulienresults.png|300px]]

Revision as of 18:49, 28 November 2011

Being edited by Rui Correia


Citation

Philipp Koehn, Franz Josef Och, and Daniel Marcu. 2003. Statistical phrase-based translation. In Proceedings of HLT-NAACL 2003, pages 127–133. [1]

Summary

In this paper the authors propose a new framework that aims at explaining and understanding why phrase-based models in Machine Translation outperform word-based models.

Within this framework (phrase-based translation model and decoding algorithm) the authors carry experiments that explore three different methods for learning phrase translation (based on word alignments, on syntactic information and "pure" phrase alignments). Additionally the authors also explore phrase length, lexical weighting, and the impact of different language pairs in the overall BLEU score.

The results confirm the already proved hypotheses that phrase translation achieve better performance than word-based methods, adding that three-word phrase are sufficient to outperform the traditional methods. Moreover, the authors conclude that lexical weighting of phrase translation boost results, and that syntactic considerations, on the other hand, hinder the results.

Evaluation Framework

The phrase translation model used in the proposed framework is based on the noisy channel model. The best English output sentence given a foreign input sentence is given by:

where:

  • is the translation model (see below);
  • is a trigram language model;
  • and, is a factor that calibrates the output length (\omega > 1, biasing longer output).


The translation model can be decomposed into:

where:

  • is a sequence of segmented from the input sentence ;
  • is a probability distribution that models the phrase translation;
  • and, is a relative distortion probability distribution between the start position of the foreign phrase that was translated into the th English phrase () and the end position of the foreign phrase translated into the th English phrase ().

The decoder that was adopted in the framework employs a Beam Search algorithm.

Methods for Learning Phrase Translation

In this work the authors compare three methods to build phrase translation probability tables. The first one builds the phrase alignments using word alignment information, i.e., all the phrase pairs that are considered must be consistent with the word alignments.

The second method explored act as a filter to the previous set of alignments, restricting possible phrases to syntactically correct ones.

Finally, the last method takes the Marcus and Wong, EMNLP 2002 approach, learning phrase-level alignments directly from the parallell corpora.

Experimental Results

The authors used the EUROPARL for the pair German-English.

Lacostejulienresults.png