10-601 Logistic Regression
From Cohen Courses
This a lecture used in the Syllabus for Machine Learning 10-601
Slides
Readings
- William's notes on SGD (for 10605)
- Lazy Sparse Stochastic Gradient Descent for Regularized Multinomial Logistic Regression, Carpenter, Bob. 2008. See also his blog post on logistic regression.
Assignment
- Implement logistic regression and apply it to a couple of datasets, using a gradient descent to optimize. Experiment by changing the regularization parameter. (Details to be posted later.)
What You Should Know Afterward
- How to implement logistic regression.
- Why regularization matters.
- How logistic regression and naive Bayes are similar and different.
- What "overfitting" is, and why optimizing performance on a training set does not necessarily lead to good performance on a test set.