10-601 Topic Models

From Cohen Courses
Revision as of 10:05, 14 November 2013 by Wcohen (talk | contribs) (Created page with "This a lecture used in the Syllabus for Machine Learning 10-601 === Slides === * [http://www.cs.cmu.edu/~wcohen/10-601/topic-models.pptx Slides in PowerPoint]. === Read...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigationJump to search

This a lecture used in the Syllabus for Machine Learning 10-601

Slides

Readings

Summary

You should know:

  • d-separation, or more generally, how to determine if two variables are conditionally independent in a directed model, and what that means.
  • what "explaining away" refers to.
  • what Gibbs sampling is, and how it can be used for inference in a directed graphical model.
  • what the graphical models are which are associated with supervised naive Bayes, unsupervised naive Bayes, and LDA.