Difference between revisions of "10-601 Matrix Factorization"

From Cohen Courses
Jump to navigationJump to search
Line 3: Line 3:
 
=== Slides ===
 
=== Slides ===
  
*  [http://www.cs.cmu.edu/~wcohen/10-601/601-pca.pptx  Slides in Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-601/601-pca.pdf  Slides in PDF].
+
*  [http://www.cs.cmu.edu/~wcohen/10-601/cf.pptx  Slides in Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-601/cf.pdf  Slides in PDF].
  
 
=== Readings ===
 
=== Readings ===

Revision as of 13:26, 13 April 2016

This a lecture used in the Syllabus for Machine Learning 10-601B in Spring 2016

Slides

Readings

Summary

You should know:

  • What loss function and constraints are associated with PCA - i.e., what the "PCA optimization problem" is.
  • How to interpret the low-dimensional embedding of instances, and the "prototypes" produced by PCA and MF techniques.
  • How to interpret the prototypes in the case of dimension reduction for images.