Difference between revisions of "10-601B Perceptrons and Large Margin"

From Cohen Courses
Jump to navigationJump to search
(Created page with "This a lecture used in the Syllabus for Machine Learning 10-601B in Spring 2016 === Slides === * === Readings === * [http://www.cs.cmu.edu/~wcohen/10-601/vp-notes/vp.p...")
 
Line 16: Line 16:
 
* How to implement the voted perceptron.
 
* How to implement the voted perceptron.
 
* The definition of a mistake bound, and a margin.
 
* The definition of a mistake bound, and a margin.
 +
 +
* The definitions of, and intuitions behind, these concepts:
 +
** The ''margin'' of a classifier relative to a dataset.
 +
** What a ''constrained optimization problem'' is.
 +
** The ''primal form'' of the SVM optimization problem.
 +
** The ''dual form'' of the SVM optimization problem.
 +
** What a ''support vector'' is.
 +
** What a ''kernel function'' is.
 +
** What ''slack variables'' are and why and when they are used in SVMs.
 +
* How to explain the different parts (constraints, optimization criteria) of the primal and dual forms for the SVM.
 +
* How the perceptron and SVM are similar and different.

Revision as of 09:15, 12 January 2016

This a lecture used in the Syllabus for Machine Learning 10-601B in Spring 2016

Slides

Readings

What You Should Know Afterward

  • The difference between an on-line and batch algorithm.
  • How to implement the voted perceptron.
  • The definition of a mistake bound, and a margin.
  • The definitions of, and intuitions behind, these concepts:
    • The margin of a classifier relative to a dataset.
    • What a constrained optimization problem is.
    • The primal form of the SVM optimization problem.
    • The dual form of the SVM optimization problem.
    • What a support vector is.
    • What a kernel function is.
    • What slack variables are and why and when they are used in SVMs.
  • How to explain the different parts (constraints, optimization criteria) of the primal and dual forms for the SVM.
  • How the perceptron and SVM are similar and different.