recurrent neural networks

2020-02-27 58浏览

  • 1.Recurrent Neural Networks Presented by Chris Foster
  • 2.RNN Hidden States A recurrent neural network node contains a hidden state h
  • 3.RNN Hidden States In practice, we also often have output y
  • 4.Back Propogation Through Time (BPTT) Apply the backpropogation algorithm to the unrolled computational graph
  • 5.BPTT is comparitively very expensive to perform Often, you can only hold limited steps in memory Training occurs in "batches" of the dataset Training is more difficult to parallelize h is initialized to 0's or the most recent h
  • 6.RNN Architectures One-to-one Vanilla processing modeEx:Image classification
  • 7.RNN Architectures One-to-many Sequence outputEx:Image captionining
  • 8.RNN Architectures Many-to-one Sequence inputEx:Sentiment analysis
  • 9.RNN Architectures Many-to-many Sequence input/outputEx:Machine translation
  • 10.RNN Architectures Many-to-many Synced sequence input/outputEx:Video labelling
  • 11.Other Network Types
  • 12.Applications of RNN's
  • 13.Applications of RNN's
  • 14.Applications of RNN's
  • 15.Finding Interpretable Cells
  • 16.Vanishing Gradient Problem
  • 17.Vanishing Gradient Problem Gradients in earlier layers are unstable This is noticable even in very deep FFN's An RNN, as we know, is like a very deep FFN The problem is made worse due to a static W
  • 18.Gradient Clipping
  • 19.Recall:Resnet Gradient Shortcuts
  • 20.The LSTM Network
  • 21.The LSTM Network
  • 22.The LSTM Network
  • 23.The LSTM Network
  • 24.The LSTM Network
  • 25.The LSTM Network
  • 26.RNN vs LSTM
  • 27.RNN vs LSTM
  • 28.Gated Recurrent Unit
  • 29.Summary RNNs allow a lot of flexibility in architecture Vanilla RNNs are simple but don't work very well Practical networks should use a LSTM/GRU Backward flow of gradients can explode or vanish
  • 30.Thanks!
  • 31.Sources Deep Learning, Chapter 10 Understanding LSTM Networks Recurrent Neural Networks Tutorial The Unreasonable Effectiveness of Recurrent Neural Networks CS231N Lecture 10 - Recurrent Neural Networks, Image Captioning, LSTM