Difference between revisions of "10-601B Clustering"

From Cohen Courses
Jump to navigationJump to search
Line 3: Line 3:
 
=== Slides ===
 
=== Slides ===
  
* ...
+
* [http://curtis.ml.cmu.edu/w/courses/images/5/5d/Clustering.pdf Slides in pdf ]
  
 
=== Readings ===
 
=== Readings ===
  
Mitchell 6.12 -  a nice description of EM and k-means.
+
* Murphy 25.5
  
 
=== What You Should Know Afterward ===
 
=== What You Should Know Afterward ===
  
You should know how to implement these methods, and what their relative advantages and disadvantages are.
+
** Partitional Clustering. k-means and k-means ++
* Overview of clustering
+
* Lloyd’s method
* Distance functions and similarity measures and their impact
+
* Initialization techniques (random, furthest traversal, k-means++)
* k-means algorithms (Lloyd's method, k-means++)
+
 
* Partitional clustering
+
 
* Hierarchical clustering
+
* Hierarchical Clustering.
* How to chose k and what is the impact of large and small k's
+
** Single linkage, Complete linkage
<!-- * Differences between GM and K-means -->
 

Revision as of 20:44, 2 March 2016

This a pair of lectures used in the Syllabus for Machine Learning 10-601B in Spring 2016.

Slides

Readings

  • Murphy 25.5

What You Should Know Afterward

    • Partitional Clustering. k-means and k-means ++
  • Lloyd’s method
  • Initialization techniques (random, furthest traversal, k-means++)


  • Hierarchical Clustering.
    • Single linkage, Complete linkage