Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Near Memory Compute for AI Inferencing - Optimizing Data Center Design and TCO

Open Compute Project via YouTube

Overview

Explore a 21-minute technical presentation that delves into innovative solutions for AI inferencing data center challenges, focusing on Near-Memory Compute (NMC) technology. Learn how low-cost remote memory connected through low-latency interconnects can optimize inferencing operations. Discover the benefits of offloading specific inferencing tasks to smaller cores positioned near remote memory, supported by simulation data demonstrating reduced execution latency. Understand the potential impact on Total Cost of Ownership (TCO) reduction in inferencing data centers through the implementation of cost-effective remote memory pools. Gain insights into forward-thinking data center design principles that prioritize both sustainability and operational efficiency.

Syllabus

Near Memory Compute for AI Inferencing

Taught by

Open Compute Project

Reviews

Start your review of Near Memory Compute for AI Inferencing - Optimizing Data Center Design and TCO

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.