Neural Nets for NLP - Recurrent Networks for Sentence or Language Modeling
Graham Neubig via YouTube
Overview
Syllabus
Intro
Why Model Sentence Pairs?
Siamese Network (Bromley et al. 1993)
Convolutional Matching Model (Hu et al. 2014) • Concatenate sentences into a 30 tensor and perform convolution
Convolutional Features + Matrix-based Pooling in and Schutze 2015
NLP and Sequential Data
Long-distance Dependencies in Language
Can be Complicated!
Recurrent Neural Networks (Elman 1990)
Unrolling in Time • What does processing a sequence look like?
What Can RNNs Do?
Representing Sentences
e.g. Language Modeling
RNNLM Example: Loss Calculation and State Update
Vanishing Gradient • Gradients decrease as they get pushed back
LSTM Structure
What can LSTMs Learn? (2) (Shi et al. 2016, Radford et al. 2017) Count length of sentence
Handling Long Sequences
Taught by
Graham Neubig