Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

CNCF [Cloud Native Computing Foundation]

Accelerator Chaining for Efficient AI/ML Workloads in Kubernetes

CNCF [Cloud Native Computing Foundation] via YouTube

Overview

Explore how to efficiently handle large AI/ML workloads in Kubernetes using accelerator chaining in this 37-minute conference talk from CNCF. Learn about extending Kubernetes with Custom Resource-based architecture and operators to orchestrate and configure device chains like FPGAs, GPUs, TPUs, and ASICs. Discover the benefits of direct data transfer between devices, including reduced memory copies, decreased CPU overheads, and lower latency. Gain insights into deploying these workloads easily and understand future developments with Dynamic Resource Allocation (DRA) support and CNI extensions. Presented by Sampath Priyankara from Nippon Telegraph and Telephone Corporation and Masataka Sonoda from Fujitsu Limited, this talk offers valuable knowledge for optimizing AI/ML performance in Kubernetes environments.

Syllabus

Accelerators(FPGA/GPU) Chaining to Efficiently Handle Large AI/ML Workloads in K8s

Taught by

CNCF [Cloud Native Computing Foundation]

Reviews

Start your review of Accelerator Chaining for Efficient AI/ML Workloads in Kubernetes

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.