BinLu et al. ACL2011
Contents
Citation
Joint Bilinguial Sentiment Classification with Unlabeled Parallel Corpora, Bin Lu, Chenhao Tan, Claire Cardie and Benjamin K. Tsou, ACL 2011
Online version
Joint Bilingual Sentiment Classification with Unlabeled Parallel Corpora
Summary
This paper address the Sentiment analysis problem on sentence level for multiple languages. They propose to leverage parallel corpora to learn a MaxEnt-based EM model that consider both languages simultaneously under the assumption that sentiment labels for parallel sentences should be similar.
The experimented on 2 dataset: MPQA Multi-Perspective Question Answering and NTCIR-6 Opinion
Evaluation
This paper compared their method with other 3 kind of state-of-the-art baseline algorithms.
1. The first kind of baseline algorithms are training separate classifiers on different languages. For this kind, the authors used MaxEnt, SVM and Monolingual TSVM 2. The second kind of baseline is Bilingual TSVM 3. The third kind is semi-supervised learning strategy Co-training
Discussion
This paper
Related papers
In sense of multilingual sentiment analysis, there several works like:
- Paper:Learning multilingual subjective language via cross-lingual projections:[1]
- Paper:Multilingual subjectivity: Are more languages better?:[2]
- Paper:Cross-language text classification using structural correspondence learning.:[3]
In sense of semi-supervised learning, related papers include:
- Paper:Combining labeled and unlabeled data with co-training:[4]
- Paper:Text classification from labeled and unlabeled documents using EM.:[5]
Study plan
- Article:Expectation Maximization Algorithm:Expectation-maximization algorithm
- Article:Maximum Entropy Model:Maximum Entropy model
- Paper:Combining labeled and unlabeled data with co-training:[6]
- Paper:Learning multilingual subjective language via cross-lingual projections:[7]