Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion

Valence Labs via YouTube

Overview

Explore a comprehensive lecture on Diffusion Forcing, a novel training paradigm for sequence generative modeling. Delve into the concept of training a diffusion model to denoise tokens with independent per-token noise levels, combining the strengths of next-token prediction and full-sequence diffusion models. Learn about the method's ability to generate variable-length sequences, guide sampling towards desirable trajectories, and roll out continuous token sequences beyond the training horizon. Discover new sampling and guiding schemes unique to Diffusion Forcing's architecture, leading to improved performance in decision-making and planning tasks. Gain insights into the theoretical foundations of the approach, including its optimization of a variational lower bound on subsequence likelihoods. The lecture covers background information, the core principles of Diffusion Forcing, its application with causal uncertainty, and concludes with a Q&A session.

Syllabus

- Intro + Background
- Diffusion Forcing
- DF with Causal Uncertainty
- Q+A

Taught by

Valence Labs

Reviews

Start your review of Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.