Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Fine-Tuning Large Language Models with Limited Resources - NeurIPS Hacker Cup AI

Weights & Biases via YouTube

Overview

Explore the intricacies of fine-tuning large language models under resource constraints in this live session from the NeurIPS Hacker Cup AI competition. Delve into optimization techniques like gradient accumulation, activation checkpointing, and LoRa as Joe Cunnings from Meta's torchtune team shares strategies for maximizing model performance with limited hardware. Learn the importance of high-quality datasets and gain practical advice for working within a 40 GB VRAM GPU environment. Perfect for developers seeking to enhance their skills in efficient model fine-tuning and competition-ready AI development. Access the torchtune GitHub repository at https://github.com/pytorch/torchtune for additional resources and tools.

Syllabus

NeurIPS Hacker Cup AI: FineTuning

Taught by

Weights & Biases

Reviews

Start your review of Fine-Tuning Large Language Models with Limited Resources - NeurIPS Hacker Cup AI

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.