Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

LLM Alignment - Techniques for Building Human-Aligned AI

Data Science Dojo via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the cutting-edge techniques for aligning Large Language Models (LLMs) with human values and ethics in this informative webinar. Delve into the evolution of LLMs from their inception to their current sophisticated forms, and discover how advanced alignment methodologies are shaping the future of AI. Learn about crucial strategies such as Reinforcement Learning from Human Feedback (RLHF), Instruction Fine-Tuning (IFT), and Direct Preference Optimization (DPO) that are making AI systems safer and more reliable. Gain insights into the progression from early models to advanced LLMs, understand the importance of RLHF in aligning AI with human values, and explore the effectiveness of IFT and DPO in refining LLM responses. Engage in discussions about ongoing challenges and ethical considerations in AI alignment. Join Hoang Tran, Senior Research Scientist at Snorkel AI, for this hour-long session that promises to deepen your understanding of building human-aligned AI systems.

Syllabus

LLM Alignment: Techniques for Building Human-Aligned AI

Taught by

Data Science Dojo

Reviews

Start your review of LLM Alignment - Techniques for Building Human-Aligned AI

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.