Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

CMU Neural Nets for NLP 2018 - Conditioned Generation

Graham Neubig via YouTube

Overview

Explore conditioned generation in neural networks for natural language processing through this comprehensive lecture from Carnegie Mellon University's 2018 Neural Nets for NLP course. Delve into key concepts including language models, conditioned language models, and various generation techniques such as ancestral sampling and ensembling. Learn about evaluation methods like human evaluation and perplexity, and gain insights into the differences between linear and log-linear models. Discover advanced topics such as parameter averaging, ensemble distillation, and stacking. Conclude with a contrastive look at evaluating unconditioned generation, providing a well-rounded understanding of this crucial aspect of NLP.

Syllabus

Intro
Language Models Language models are generative models of text
Conditioned Language Models
Ancestral Sampling
Ensembling
Linear or Log Linear?
Parameter Averaging
Ensemble Distillation (e.g. Kim et al. 2016)
Stacking
Basic Evaluation Paradigm
Human Evaluation
Perplexity
A Contrastive Note: Evaluating Unconditioned Generation

Taught by

Graham Neubig

Reviews

Start your review of CMU Neural Nets for NLP 2018 - Conditioned Generation

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.