Difference between revisions of "10-601 Deep Learning 2"

From Cohen Courses
Jump to navigationJump to search
Line 28: Line 28:
 
** How to construct a convolution layer
 
** How to construct a convolution layer
 
** Architecture of CNN: convolution/downsampling pairs
 
** Architecture of CNN: convolution/downsampling pairs
 +
* Recurrent neural networks
 +
** When they are useful
 +
** Why they are hard to train (if trained naively)
 +
** The basic ideas used in an LSTM: forget, insert, and output gates

Revision as of 14:23, 11 April 2016

This a lecture used in the Syllabus for Machine Learning 10-601B in Spring 2016

Slides

Readings

This area is moving very fast and the textbooks are not up-to-date. Some recommended readings:

I also used some on-line visualizations in the materials for the lecture, especially the part on ConvNets.

The LSTM figures and examples I used are mostly from

Things to remember

  • Convolutional networks
    • 2-d convolution
    • How to construct a convolution layer
    • Architecture of CNN: convolution/downsampling pairs
  • Recurrent neural networks
    • When they are useful
    • Why they are hard to train (if trained naively)
    • The basic ideas used in an LSTM: forget, insert, and output gates