Difference between revisions of "10-601 Logistic Regression"

From Cohen Courses
Jump to navigationJump to search
Line 3: Line 3:
 
=== Slides ===
 
=== Slides ===
  
* [http://www.cs.cmu.edu/~wcohen/10-601/logreg.pptx Slides in Powerpoint]. ''Based on the slides I used for 10-605, they might be updated.''
+
* [http://www.cs.cmu.edu/~wcohen/10-601/logreg.pptx Slides in Powerpoint].
  
 
=== Readings ===
 
=== Readings ===

Revision as of 09:46, 18 September 2013

This a lecture used in the Syllabus for Machine Learning 10-601

Slides

Readings

  • None

Assignment

  • Implement logistic regression and apply it to a couple of datasets, using a gradient descent to optimize. Experiment by changing the regularization parameter. (Details to be posted later.)

What You Should Know Afterward

  • How to implement logistic regression.
  • Why regularization matters.
  • How logistic regression and naive Bayes are similar and different.
  • What "overfitting" is, and why optimizing performance on a training set does not necessarily lead to good performance on a test set.