Difference between revisions of "10-601 Decision Trees"
From Cohen Courses
Jump to navigationJump to search (→Slides) |
(→Slides) |
||
Line 3: | Line 3: | ||
=== Slides === | === Slides === | ||
− | * Ziv's lecture: to be posted | + | * Ziv's lecture: ''to be posted'' |
* William's lecture: [http://www.cs.cmu.edu/~wcohen/10-601/decision-trees.pptx Slides in Powerpoint]. | * William's lecture: [http://www.cs.cmu.edu/~wcohen/10-601/decision-trees.pptx Slides in Powerpoint]. | ||
Revision as of 12:40, 15 August 2014
This a lecture used in the Syllabus for Machine Learning 10-601 in Fall 2014
Slides
- Ziv's lecture: to be posted
- William's lecture: Slides in Powerpoint.
Readings
- Mitchell, Chapter 3.
What You Should Know Afterward
- What a decision tree is, and how to classify an instance using a decision tree.
- What the canonical top-down algorithm is for learning a decision tree.
- What heuristics are used for choosing a decision-tree split.
- What entropy is, and what information gain is.
- What reduced-error pruning is, and why it might improve classification performance.
- Some of the advantages and disadvantages of decision-tree learning in specific, and eager learning in general, compared to K-NN learning.