Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Direct Preference Optimization - Fine-Tuning LLMs Without Reinforcement Learning

Serrano.Academy via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the innovative Direct Preference Optimization (DPO) method for training Large Language Models in this 21-minute video tutorial. Discover how DPO offers a more effective and efficient alternative to reinforcement learning techniques. Delve into key concepts such as the Bradley-Terry Model, KL Divergence, and the Loss Function. Compare DPO with Reinforcement Learning with Human Feedback (RLHF) to understand its advantages. As the third installment in a four-part series on reinforcement learning methods for LLMs, this video provides valuable insights for those interested in advanced machine learning techniques. Access additional resources, including related videos in the series and a recommended book on machine learning, to further enhance your understanding of LLM training methodologies.

Syllabus

Introduction
RLHF vs DPO
The Bradley-Terry Model
KL Divergence
The Loss Function
Conclusion

Taught by

Serrano.Academy

Reviews

Start your review of Direct Preference Optimization - Fine-Tuning LLMs Without Reinforcement Learning

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.