Difference between revisions of "10-601 Linear Regression"
From Cohen Courses
Jump to navigationJump to search (→Slides) |
(→Slides) |
||
Line 4: | Line 4: | ||
* Ziv's lecture: [http://www.cs.cmu.edu/~zivbj/classF14/RegNew.pdf Slides in pdf]. | * Ziv's lecture: [http://www.cs.cmu.edu/~zivbj/classF14/RegNew.pdf Slides in pdf]. | ||
− | * William's lecture: [http://www.cs.cmu.edu/~wcohen/10-601/linear-regression.ppt Slides in Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-601/linear-regression.pdf in | + | * William's lecture: [http://www.cs.cmu.edu/~wcohen/10-601/linear-regression.ppt Slides in Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-601/linear-regression.pdf in PDF]. |
=== Readings === | === Readings === |
Revision as of 13:39, 17 September 2014
This a lecture used in the Syllabus for Machine Learning 10-601 in Fall 2014
Slides
- Ziv's lecture: Slides in pdf.
- William's lecture: Slides in Powerpoint, in PDF.
Readings
- Mitchell 4.1-4.3
- Optional:
- Bishop 3.1
What You Should Know Afterward
- Regression vs. classification
- Solving regression problems with 1 and 2 variables
- Ordinary least squares (OLS) solution (aka normal equations) to linear regression problems
- Gradient descent approach to linear regression
- Data transformation and its impact on the way linear regression is solved, and the expressiveness of LR models