Shmueli et. al. WWW2012

From Cohen Courses
Revision as of 02:42, 2 October 2012 by Yuchent (talk | contribs)
Jump to navigationJump to search

Citation

E Shmueli, A Kagian, Y Koren, R Lempel Care to Comment? Recommendations for Commenting on News Stories WWW2012

Summary

In this paper, the authors address the following problem, "How to recommend News Stories to recommend?"

They treat this problem as a classic recommendation problem. They use two approaches to recommend article to users. Memory based approach and Latent Factor Model.

Datasets

The author established two data sets: News Vine and Yahoo! For News Vine, the author crawled all the articles from May to September 2011, including all the tags and all the associated comments. They also crawled the friends relationships from for each commender.

And the Yahoo! data sets are constructed following the same way.

Approaches

1.Memory Based Approach. The authors defined a co-occurrence similarity function using both tag information and co-commenter information. Then rank stories using this function.

2.Latent Factor Model The authors define a K dimensional latent space for both users and stories and the product of two vectors plus a bias function as a rank function. Then the model is trained by minimizing a loss function defined later, using stochastic gradient descent.

Evaluation

The evaluation is performed as follows, the authors randomly selected pairs of stories, one from the stories the user made a commend and those stories the user didn't. Then using the measure defined above, to assign scores to pick one of the story over another, then using AUC metric to evaluate the performance.

As to the Latent Model approach, the user using one of the following metrics, to define the loss of misclassification. 1.Squared Error 2.Classification Error 3.Bayesian Personalized Ranking 4.Rank Loss

The result shows that on the Newsvine Datasets, Latent Factor model with BPR loss function performs the best, and on Yahoo! dataset the LFM+RL performs better.

Stratified AUC

The authors found some problem on the above methods have a common weakness: that is many stories are unlikely to be commend by any user. It makes more sense to distinguish between stories that has a similar popularity.

So the author put stories into different popularity bins, the stories in the same bin share the same popularity.

Then AUC is computed on each bins and averaged proportionally to the number of positive examples in each bin.

This time it shows that BPR has the best performance on Newsvine datasets while CE has the best performance on Yahoo.

My Opinion

This