Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Parameter-Efficient Fine-Tuning with LoRA - Optimizing LLMs for Local GPU Training

Discover AI via YouTube

Overview

Learn how to efficiently fine-tune Large Language Models (LLMs) on local GPUs with limited memory through a detailed 41-minute technical video exploring PEFT (Parameter-Efficient Fine-Tuning) and LoRA (Low-Rank Adaptation) methodologies. Dive deep into the mathematical foundations of weight tensor approximation using eigenvector and eigenvalue decomposition, understanding how these techniques enable minimal GPU/TPU memory requirements. Master the implementation of HuggingFace's PEFT library for transformer models across language, image, and vision applications, while exploring advanced concepts like 8-bit quantization, adapter-tuning, and optimal LoraConfig settings. Discover how combining non-trainable pre-trained weights with reduced memory footprint techniques achieves state-of-the-art benchmarks compared to traditional fine-tuning approaches for models like GPT, BLOOM, LLama, and T5. Progress through comprehensive topics including adapter transformers, weight matrices, rank decomposition, singular value decomposition, and practical LoRA configurations, concluding with performance analysis and institutional applications.

Syllabus

Intro
Use cases
Path import
Adapter transformers
What are adapters
Adapter Hub
Weight Mattresses
LoRA
Rank Decomposition
Singular Value decomposition
Low Rank Adaptation
Data Analysis
Singular Values
Performance
LoRA as an institution
LoRA configuration files
Summary

Taught by

Discover AI

Reviews

Start your review of Parameter-Efficient Fine-Tuning with LoRA - Optimizing LLMs for Local GPU Training

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.