Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Advanced NLP 2022: Modeling Long Sequences

Graham Neubig via YouTube

Overview

Explore advanced techniques for modeling long sequences in natural language processing through this comprehensive lecture from CMU's Advanced NLP course. Delve into extracting features from extended text and tackling document processing tasks. Learn about various transformer architectures including Transformer XL, Compressive Transformers, and Sparse Transformers. Examine adaptive span and sparse span approaches, as well as the Reformer model. Investigate low rank approximation and sparse attention methods. Gain insights into evaluation techniques and other relevant methodologies. Conclude with an overview of coreference models, including mention pair models and their components.

Syllabus

Introduction
NLP Tasks
Modeling Long Sequences
Separate Encoding
Selfattention Transformers
Transformer XL
Compressive Transformers
Sparse Transformers
Adaptive Span Transformers
Sparse Span Transformers
Reformer Model
Low Rank Approximation
Sparse Attention
Evaluation
Other Methods
Questions
Components of Coreference Models
Mention Pair Models
Model

Taught by

Graham Neubig

Reviews

Start your review of Advanced NLP 2022: Modeling Long Sequences

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.