Difference between revisions of "10-601 Logistic Regression"
From Cohen Courses
Jump to navigationJump to searchLine 10: | Line 10: | ||
* [http://cseweb.ucsd.edu/~elkan/250B/logreg.pdf Charles Elkan's notes on SGD] | * [http://cseweb.ucsd.edu/~elkan/250B/logreg.pdf Charles Elkan's notes on SGD] | ||
* [http://lingpipe.files.wordpress.com/2008/04/lazysgdregression.pdf Lazy Sparse Stochastic Gradient Descent for Regularized Multinomial Logistic Regression], Carpenter, Bob. 2008. See also [http://alias-i.com/lingpipe/demos/tutorial/logistic-regression/read-me.html his blog post] on logistic regression. | * [http://lingpipe.files.wordpress.com/2008/04/lazysgdregression.pdf Lazy Sparse Stochastic Gradient Descent for Regularized Multinomial Logistic Regression], Carpenter, Bob. 2008. See also [http://alias-i.com/lingpipe/demos/tutorial/logistic-regression/read-me.html his blog post] on logistic regression. | ||
− | |||
− | |||
− | |||
− | |||
=== What You Should Know Afterward === | === What You Should Know Afterward === |
Revision as of 11:48, 24 September 2013
This a lecture used in the Syllabus for Machine Learning 10-601
Slides
Readings
- William's notes on SGD (for 10605)
- Charles Elkan's notes on SGD
- Lazy Sparse Stochastic Gradient Descent for Regularized Multinomial Logistic Regression, Carpenter, Bob. 2008. See also his blog post on logistic regression.
What You Should Know Afterward
- How to implement logistic regression.
- Why regularization matters.
- How logistic regression and naive Bayes are similar and different.
- What "overfitting" is, and why optimizing performance on a training set does not necessarily lead to good performance on a test set.