Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

LLM Efficient Inference in CPUs and Intel GPUs - Intel Neural Speed

The Machine Learning Engineer via YouTube

Overview

Explore efficient inference techniques for Large Language Models (LLMs) on CPUs and Intel GPUs using Intel Neural Speed in this 30-minute video. Dive into the performance capabilities of Intel Extension for Transformers and gain practical insights through provided Jupyter notebooks. Learn how to optimize LLM inference for data science and machine learning applications, leveraging Intel's hardware-specific solutions. Access accompanying resources, including a Medium article and GitHub repositories, to deepen your understanding and implement the techniques discussed.

Syllabus

LLM Efficient Inference In CPUs and Intel GPUs. Intel Neural Speed #datascience #machinelearning

Taught by

The Machine Learning Engineer

Reviews

Start your review of LLM Efficient Inference in CPUs and Intel GPUs - Intel Neural Speed

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.