Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

FwdLLM: Efficient Federated Finetuning of Large Language Models with Perturbed Inferences

USENIX via YouTube

Overview

Explore an innovative approach to federated fine-tuning of Large Language Models (LLMs) in this 20-minute conference talk from USENIX ATC '24. Dive into FwdLLM, a novel protocol designed to enhance efficiency in Federated Learning (FL) for LLMs on mobile devices. Learn how the researchers from Beijing University of Posts and Telecommunications address the challenge of balancing LLM complexity with mobile resource constraints. Discover the key components of FwdFL, including backpropagation-free training methods, adaptive computational load allocation, and discriminative sampling of perturbed predictions. Gain insights into the significant advantages of this approach, such as faster convergence and reduced memory footprint, and understand how it enables federated billion-parameter LLMs on commercial off-the-shelf mobile devices for the first time.

Syllabus

USENIX ATC '24 - FwdLLM: Efficient Federated Finetuning of Large Language Models with Perturbed...

Taught by

USENIX

Reviews

Start your review of FwdLLM: Efficient Federated Finetuning of Large Language Models with Perturbed Inferences

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.