Difference between revisions of "10-601 Logistic Regression"

From Cohen Courses
Jump to navigationJump to search
Line 10: Line 10:
 
* [http://cseweb.ucsd.edu/~elkan/250B/logreg.pdf Charles Elkan's notes on SGD]
 
* [http://cseweb.ucsd.edu/~elkan/250B/logreg.pdf Charles Elkan's notes on SGD]
 
* [http://lingpipe.files.wordpress.com/2008/04/lazysgdregression.pdf Lazy Sparse Stochastic Gradient Descent for Regularized Multinomial Logistic Regression], Carpenter, Bob. 2008. See also [http://alias-i.com/lingpipe/demos/tutorial/logistic-regression/read-me.html his blog post] on logistic regression.
 
* [http://lingpipe.files.wordpress.com/2008/04/lazysgdregression.pdf Lazy Sparse Stochastic Gradient Descent for Regularized Multinomial Logistic Regression], Carpenter, Bob. 2008. See also [http://alias-i.com/lingpipe/demos/tutorial/logistic-regression/read-me.html his blog post] on logistic regression.
 
=== Assignment ===
 
 
* Implement logistic regression and apply it to a couple of datasets, using a gradient descent to optimize. Experiment by changing the regularization parameter and learning rate.
 
  
 
=== What You Should Know Afterward ===
 
=== What You Should Know Afterward ===

Revision as of 11:48, 24 September 2013

This a lecture used in the Syllabus for Machine Learning 10-601

Slides

Readings

What You Should Know Afterward

  • How to implement logistic regression.
  • Why regularization matters.
  • How logistic regression and naive Bayes are similar and different.
  • What "overfitting" is, and why optimizing performance on a training set does not necessarily lead to good performance on a test set.