Difference between revisions of "Bartlett et al., ACL-HLT 2008. Automatic Syllabification with Structured SVMs for Letter-to-Phoneme Conversion"

From Cohen Courses
Jump to navigationJump to search
Line 11: Line 11:
  
 
=== Structured SVMs ===
 
=== Structured SVMs ===
[[UsesMethod::Structured SVMs | Structured SVMs]] [[Tsochantaridis,_Joachims_,_Support_vector_machine_learning_for_interdependent_and_structured_output_spaces_2004 (Tsochantaridis et al., 2004)]] is a large-margin training method that is used for predicting structured outputs such as tag sequences. The method described in this paper uses structured SVMs that learn tag sequences from training data and perform structured output prediction by finding the highest scoring tag sequence using the Viterbi algorithm. Hence, the decoding problem in [http://en.wikipedia.org/wiki/Structured_SVM structured SVMs] resembles that of an [[HMM|HMM]].
+
[[UsesMethod::Structured SVMs | Structured SVMs]] [[Tsochantaridis,_Joachims_,_Support_vector_machine_learning_for_interdependent_and_structured_output_spaces_2004 | (Tsochantaridis et al., 2004)]] is a large-margin training method that is used for predicting structured outputs such as tag sequences. The method described in this paper uses structured SVMs that learn tag sequences from training data and perform structured output prediction by finding the highest scoring tag sequence using the Viterbi algorithm. Hence, the decoding problem in [http://en.wikipedia.org/wiki/Structured_SVM structured SVMs] resembles that of an [[HMM|HMM]].
  
 
Training structured SVMs is viewed as a multi-class classification problem. For a given training instance <math>x_i</math>, a correct tag sequence <math>y_i</math> is drawn from a set of possible tag sequences <math>Y_i</math>. Each input sequence <math>x</math> has a feature space representation <math>\Psi (x,y)</math> to represent a candidate output sequence <math>y</math>.
 
Training structured SVMs is viewed as a multi-class classification problem. For a given training instance <math>x_i</math>, a correct tag sequence <math>y_i</math> is drawn from a set of possible tag sequences <math>Y_i</math>. Each input sequence <math>x</math> has a feature space representation <math>\Psi (x,y)</math> to represent a candidate output sequence <math>y</math>.

Revision as of 14:12, 28 September 2011

Citation

Susan Bartlett, Grzegorz Kondrak and Colin Cherry. 2008. Automatic syllabification with structured SVMs for letter-to-phoneme conversion. In Proceedings of ACL-08: HLT, 2008, pp. 568–576.

Online Version

Automatic syllabification with structured SVMs for letter-to-phoneme conversion.

Summary

This paper describes one of the first successful attempts at integrating automatic syllabification into a letter-to-phoneme conversion system using structured SVMs. The authors obtain substantial improvements in reducing automatic syllabification error rate (measured in WER) against the then state-of-the-art approach. The authors model the problem as an orthographic syllabification task as opposed to phonological syllabification. They treat it as a sequence tagging problem and define new tagging schemes. The method is applied to languages such as German and Dutch, in addition to English.

Method

Structured SVMs

Structured SVMs (Tsochantaridis et al., 2004) is a large-margin training method that is used for predicting structured outputs such as tag sequences. The method described in this paper uses structured SVMs that learn tag sequences from training data and perform structured output prediction by finding the highest scoring tag sequence using the Viterbi algorithm. Hence, the decoding problem in structured SVMs resembles that of an HMM.

Training structured SVMs is viewed as a multi-class classification problem. For a given training instance , a correct tag sequence is drawn from a set of possible tag sequences . Each input sequence has a feature space representation to represent a candidate output sequence .

The training process works in a manner that weights a correctly classified input sequence more than an incorrectly classified input sequence using a weight vector . It tries to the maximize the margin between the correct and incorrect sequence. Mathematically, the relationship can be expressed as:


The model then predicts the highest scoring sequence as follows,


The search for the highest confidence score is performed by using the Viterbi algorithm. The set of negative or incorrect sequences in is potentially exponential with respect to . Hence, structured SVMs address this problem by limiting the search space for potential output sequences with the help of an iterative online learning approach. This approach is outlined as follows:

1. Find out the least scoring or most damaging incorrect sequence according to the current weight vector .

2. Add to an iteratively increasing list of of incorrect sequences.

3. Update the weight vector according to the objective function using the partial sets in place of .

4. Go to next training instance and loop from 1 until convergence.

Tagging Scheme

Experiments and Results

Related Papers