Difference between revisions of "10-601 Ensembles 1"

From Cohen Courses
Jump to navigationJump to search
 
(3 intermediate revisions by 2 users not shown)
Line 1: Line 1:
 +
This a lecture used in the [[Syllabus for Machine Learning 10-601 in Fall 2014]]
 +
 
=== Slides ===
 
=== Slides ===
 +
* [http://www.cs.cmu.edu/~zivbj/classF14/boosting.pdf Slides in pdf].
  
 
* [http://www.cs.cmu.edu/~wcohen/10-601/ensembles1.ppt Slides in PowerPoint].
 
* [http://www.cs.cmu.edu/~wcohen/10-601/ensembles1.ppt Slides in PowerPoint].
Line 11: Line 14:
  
 
You should know how to implement  these ensemble methods, and what their relative advantages and disadvantages are:
 
You should know how to implement  these ensemble methods, and what their relative advantages and disadvantages are:
* Bagging
+
* (Ziv - not sure if can do in one lecture if we do boosting) Bagging
 
* Boosting
 
* Boosting
 
* Stacking
 
* Stacking
 
* Multilevel Stacking
 
* Multilevel Stacking
* The "bucket of models" classifier
+
* (Ziv - not sure if I will do this) The "bucket of models" classifier
 +
* Random forest

Latest revision as of 07:53, 22 October 2014

This a lecture used in the Syllabus for Machine Learning 10-601 in Fall 2014

Slides

Readings

Summary

You should know how to implement these ensemble methods, and what their relative advantages and disadvantages are:

  • (Ziv - not sure if can do in one lecture if we do boosting) Bagging
  • Boosting
  • Stacking
  • Multilevel Stacking
  • (Ziv - not sure if I will do this) The "bucket of models" classifier
  • Random forest