Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

What Can Storage Do for AI? - Optimizing NVMe Storage for Large Language Models

SNIAVideo via YouTube

Overview

Explore a 32-minute conference talk from SNIA Storage Developer Conference 2024 examining the crucial role of flash storage and PCIe/NVMe in supporting AI applications across different scales. Dive into how NVMe storage can enhance both training and inference deployments, from large data centers to edge devices. Learn about the specific requirements for enabling NVMe offload in generative AI models through practical examples using the Microsoft Deep Speed library. Understand the optimization techniques and improvements needed in NVMe storage to achieve better LLM inference metrics. Presented by industry experts from Micron Technology, gain insights into democratizing AI training and inference at scale, the technical requirements for NVMe offload of LLMs, and opportunities for enhancing LLM inference performance through NVMe flash storage solutions.

Syllabus

SNIA SDC 2024 - What Can Storage Do for AI?

Taught by

SNIAVideo

Reviews

Start your review of What Can Storage Do for AI? - Optimizing NVMe Storage for Large Language Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.