Difference between revisions of "Hopkins and May, EMNLP 2011. Tuning as Ranking"

From Cohen Courses
Jump to navigationJump to search
Line 14: Line 14:
  
 
=== Approach ===
 
=== Approach ===
 +
The goal of tuning is to learn a weight vector <math> \textbf {w} </math> such that <math> H_{w}(p) </math> assigns a high score to good translations
 +
and a low score to bad translations. <math> H_{w}(p) </math> is given by:
 +
 +
<math>
 +
H_{w}(p) = \sum_{i \in I} h_{w}(i, p(i))
 +
</math>
 +
 +
where, the candidate translations are represented in the following form, <math> h_{w}(i, p(i)) = \textbf{w} \cdot \textbf {x}(i,j) </math>
  
 
== Experiments and Results ==
 
== Experiments and Results ==

Revision as of 19:23, 29 November 2011

Citation

Mark Hopkins and Jonathan May. 2011. Tuning as Ranking. In Proceedings of EMNLP-2011.

Online Version

Tuning as Ranking

Summary

This paper presents a simple and scalable method for statistical machine translation parameter tuning based on the pairwise approach to ranking. This pairwise ranking optimization (PRO) method has advantages over MERT Och, 2003 as it is not limited to a handful of parameters and can easily handle systems with thousands of features. In addition, unlike recent approaches built upon the MIRA algorithm of Crammer and Singer, 2003 (Watanabe et al., 2007), PRO is easy to implement.

Method

Although, MERT is well-understood, easy to implement, and runs quickly, it can behave erratically and would not scale beyond a handful of features. This is a major bottleneck towards working with richer feature representations and structure.

Hence, the authors propose a simpler approach than MIRA, to tuning that similarly scales to high-dimensional feature spaces. Tuning is treated as a ranking problem (Chen et al., 2009), where the explicit goal is to learn to correctly rank candidate translations. The authors describe a pairwise approach to ranking, in which the ranking problem is reduced to the binary classification task of deciding between candidate translation pairs.

Approach

The goal of tuning is to learn a weight vector such that assigns a high score to good translations and a low score to bad translations. is given by:

where, the candidate translations are represented in the following form,

Experiments and Results

Dataset

Results

Related Papers