Difference between revisions of "Class meeting for 10-605 Scalable PageRank"

From Cohen Courses
Jump to navigationJump to search
 
Line 1: Line 1:
This is one of the class meetings on the [[Syllabus for Machine Learning with Large Datasets 10-605 in Fall 2015|schedule]] for the course [[Machine Learning with Large Datasets 10-605 in Fall_2015]].
+
This is one of the class meetings on the [[Syllabus for Machine Learning with Large Datasets 10-605 in Fall 2016|schedule]] for the course [[Machine Learning with Large Datasets 10-605 in Fall_2016]].
  
 
=== Slides ===
 
=== Slides ===

Latest revision as of 13:31, 10 August 2016

This is one of the class meetings on the schedule for the course Machine Learning with Large Datasets 10-605 in Fall_2016.

Slides

Quiz

Readings

Key things to remember

  • How to implement graph algorithms like PageRank by streaming through a graph, under various conditions:
    • Vertex weights fit in memory
    • Vertex weights do not fit in memory
  • The meaning of various graph statistics: degree distribution, clustering coefficient, ...
  • Why sampling from a graph is non-trivial if you want to preserve properties of the graph like
    • Degree distribution
    • Homophily as measured by clustering coefficient,
  • What local graph partitioning is and how the PageRank-Nibble algorithm, together with sweeps to optimize conductance, can be used to approximately solve it.
  • The implications of the analysis of PageRank-Nibble.