Develop, Build and Deploy Lightweight AI Services with WebAssembly
CNCF [Cloud Native Computing Foundation] via YouTube
Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the development, building, and deployment of lightweight AI services using WebAssembly in this comprehensive tutorial. Learn how to create efficient PyTorch and TensorFlow inference functions, extend large language models like ChatGPT using Rust, and deploy AI inference functions with Kubernetes. Discover the advantages of WebAssembly as a powerful alternative runtime for AI inference workloads, offering lightweight, fast, and cloud-safe solutions. Gain hands-on experience in creating Wasm-based AI services that address the challenges of resource-intensive inference demands, particularly in the context of large language models. Understand the architecture and fundamentals of WebAssembly-based AI services, and explore how they compare to traditional Python-based approaches in terms of size, speed, and resource consumption.
Syllabus
Develop, Build & Deploy Lightweight AI Services with WebAssembly - Kinfey Lo, Vivian Hu, Jianbai Ye
Taught by
CNCF [Cloud Native Computing Foundation]