Difference between revisions of "Compare Link Propagation Papers"

From Cohen Courses
Jump to navigationJump to search
(Created page with '== Papers == == Method == == Dataset Used == == Problem == == Big Idea == == Other ==')
 
Line 1: Line 1:
 
== Papers ==
 
== Papers ==
 
+
[[ Link propagation: A fast semi-supervised learning algorithm for link prediction ]], Kashima et al., SDM 2009
 +
[[ Fast and scalable algorithms for semi-supervised link prediction on static and dynamic graphs ]], Raymond et al. ECML-PKDD 2010.
  
 
== Method ==
 
== Method ==
 +
The first paper introduces the algorithm of "Link Propagation" which is a means of doing link prediction between nodes of a Graph, using the eigen-decomposition of its Laplacian.
 +
The second paper is an extension fo the first, in the sense that it builds up on the previous algoritm, introducing an approximate version thereof which operates faster and its performance is on par with the original one. The approximate algorithm relies on Matrix Decompositions, such as the Cholesky decomposition/factorization, and in particular
 +
its main idea is the approximation of the Laplacian of the Graph in a lower rank.
  
 +
== Dataset Used ==
 +
The first paper uses a relatively large variety of different datasets, i.e.:
  
== Dataset Used ==
+
* [[UsesDataset::KEGG/PATHWAY database]]
 +
*A dataset derived from the following paper
 +
[[paper::C. von Mering, R. Krause, B. Snel, M. Cornell, S. Olivier,S. Fields, and P. Bork. Comparative assessment of large-scale data sets of protein-protein interactions. Nature, 417:399–403, 2002]]
 +
*A social network dataset: [http://ai.stanford.edu/~gal/data.html here]
  
 +
The second paper uses multiple versions (of varying size) of the [http://webspam.lip6.fr/wiki/pmwiki.php?n=Main.PhaseIITrainningCorpora Web Spam Challenge] dataset.
  
 
== Problem ==
 
== Problem ==
 +
The problem in the first paper is merely the one of predicting missing links between nodes.
  
 +
However, this is not the main focus of the second paper, even though it builds up on the same algorithm. The second paper is merely concerned with scaling up "Link Propagation" by introducing approximation in the computations, therefore it is more algorithmic oriented.
  
 
== Big Idea ==
 
== Big Idea ==

Revision as of 00:45, 5 November 2012

Papers

Link propagation: A fast semi-supervised learning algorithm for link prediction , Kashima et al., SDM 2009 Fast and scalable algorithms for semi-supervised link prediction on static and dynamic graphs , Raymond et al. ECML-PKDD 2010.

Method

The first paper introduces the algorithm of "Link Propagation" which is a means of doing link prediction between nodes of a Graph, using the eigen-decomposition of its Laplacian. The second paper is an extension fo the first, in the sense that it builds up on the previous algoritm, introducing an approximate version thereof which operates faster and its performance is on par with the original one. The approximate algorithm relies on Matrix Decompositions, such as the Cholesky decomposition/factorization, and in particular its main idea is the approximation of the Laplacian of the Graph in a lower rank.

Dataset Used

The first paper uses a relatively large variety of different datasets, i.e.:

C. von Mering, R. Krause, B. Snel, M. Cornell, S. Olivier,S. Fields, and P. Bork. Comparative assessment of large-scale data sets of protein-protein interactions. Nature, 417:399–403, 2002

  • A social network dataset: here

The second paper uses multiple versions (of varying size) of the Web Spam Challenge dataset.

Problem

The problem in the first paper is merely the one of predicting missing links between nodes.

However, this is not the main focus of the second paper, even though it builds up on the same algorithm. The second paper is merely concerned with scaling up "Link Propagation" by introducing approximation in the computations, therefore it is more algorithmic oriented.

Big Idea

Other