Outline of the lecture
This lecture introduces you sequence models. The goal is for you to
learn about:
Recurrent neural networks
The vanishing and exploding gradients problem
Long-short term memory (LSTM) networks
Applications of LSTM networks
Language models
Translation
Caption generation
Program execution
A simple recurrent neural network
[Alex Graves]
Vanishing gradient problem
[Yoshua Bengio et al]
Vanishing gradient problem
Simple solution
LSTM
[Alex Graves]
LSTM
Entry-wise multiplication layer
LSTM cell in Torch
LSTM column in Torch
LSTMs for sequence to sequence prediction
[Ilya Sutskever et al]
LSTMs for sequence to sequence prediction
Learning to parse
[Oriol Vinyals et al]
Learning to execute
[Wojciech Zaremba and Ilya Sutskever]
Video prediction
Hand-writing recognition and synthesis
[Alex Graves]
Neural Turing Machine (NTM)
[Alex Graves, Greg Wayne, Ivo Danihelka]
Neural Turing Machine (NTM)
Neural Turing Machine (NTM)
Translation with alignment (Bahdanau et al)
Show, attend and tell
[Kelvin Xu et al, 2015]
Show, attend and tell