Difference between revisions of "Class meeting for 10-405 Deep Learning"

From Cohen Courses
Jump to navigationJump to search
(Created page with "This is one of the class meetings on the schedule for the course Machine Learning with Large Data...")
 
 
(6 intermediate revisions by the same user not shown)
Line 5: Line 5:
 
* Lecture 1: [http://www.cs.cmu.edu/~wcohen/10-405/deep-1.pptx Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-405/deep-1.pdf PDF].
 
* Lecture 1: [http://www.cs.cmu.edu/~wcohen/10-405/deep-1.pptx Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-405/deep-1.pdf PDF].
  
* Lecture 2: [http://www.cs.cmu.edu/~wcohen/10-405/deep-2.pptx Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-405/deep-2.pdf PDF].
+
* Lecture 2: [http://www.cs.cmu.edu/~wcohen/10-405/deep-2.pptx Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-405/deep-2.pdf PDF]  
  
* Lecture 3: [http://www.cs.cmu.edu/~wcohen/10-405/deep-3.pptx Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-405/deep-3.pdf PDF].
+
* Lecture 3: [http://www.cs.cmu.edu/~wcohen/10-405/deep-3.pptx Powerpoint], [http://www.cs.cmu.edu/~wcohen/10-405/deep-3.pdf PDF] (draft)
  
 
=== Quizzes ===
 
=== Quizzes ===
  
* [https://qna.cs.cmu.edu/#/pages/view/75 Quiz for lecture 1]
+
 
* [https://qna.cs.cmu.edu/#/pages/view/79 Quiz for lecture 2]
+
* [https://qna.cs.cmu.edu/#/pages/view/79 Quiz for lecture 1]
 +
 
 +
* [https://qna.cs.cmu.edu/#/pages/view/246 Quiz for lecture 2]
 +
 
 
* [https://qna.cs.cmu.edu/#/pages/view/212 Quiz for lecture 3]
 
* [https://qna.cs.cmu.edu/#/pages/view/212 Quiz for lecture 3]
  

Latest revision as of 12:04, 26 March 2018

This is one of the class meetings on the schedule for the course Machine Learning with Large Datasets 10-405 in Spring 2018.

Slides

Quizzes

Sample code

Readings

Things to remember

  • The underlying reasons deep networks are hard to train
  • Exploding/vanishing gradients
  • Saturation
  • The importance of key recent advances in neural networks:
  • Matrix operations and GPU training
  • ReLU, cross-entropy, softmax
  • How backprop can be generalized to a sequence of assignment operations (autodiff)
    • Wengert lists
    • How to evaluate and differentiate a Wengert list
  • Common architectures
    • Multi-layer perceptron
    • Recursive NNs (RNNS) and Long/short term memory networks (LSTMs)
    • Convolutional Networks (CNNs)