Emergent Linguistic Structure in Deep Contextual Neural Word Representations - Chris Manning
Institute for Advanced Study via YouTube
Overview
Syllabus
Intro
1. Language Modeling
Enlightenment era neural language models (NLMs) 1. Solve curse of dimensionality by sharing of statistical strength via
Recurrent models with (self-)attention
Self-attention in masked sequence model
SQUAD Question
What do BERT attention heads do?
There's a coreference head (!)
Distance metrics unify trees and vectors
Finding trees in vector spaces
Taught by
Institute for Advanced Study