Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Extending PyTorch for Custom Compiler Targets - Accelerator Hardware Integration

Toronto Machine Learning Series (TMLS) via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn how to annotate PyTorch code with custom metadata for optimizing model inference on specialized hardware accelerators in this conference talk from the Toronto Machine Learning Series. Discover techniques presented by Groq Compiler Engineer Arash Taheri-Dezfouli for injecting and preserving arbitrary information in PyTorch graphs to improve performance on Language Processing Units (LPUs). Explore methods for maintaining graph semantics while adjusting workload mapping, implementing custom data types, and persisting precision information through PyTorch's compilation pipeline. Master generalizable approaches for annotating PyTorch models at various granularities to maximize inference efficiency on custom hardware targets, working around limitations in standard graph export systems like TorchScript, ONNX, and torch.compile.

Syllabus

Extending PyTorch for Custom Compiler Targets

Taught by

Toronto Machine Learning Series (TMLS)

Reviews

Start your review of Extending PyTorch for Custom Compiler Targets - Accelerator Hardware Integration

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.