Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Stanford University

Petascale Deep Learning on a Single Chip

Stanford University via YouTube

Overview

Explore a Stanford University seminar on petascale deep learning processor architecture, featuring Tapabrata Ghosh from Vathys.ai. Delve into innovative strategies for reducing data movement and improving efficiency in deep learning processors, including circuit-level innovations and comparisons with designs like Google's TPU. Learn about Vathys' approach to addressing computational bottlenecks, scalability for next-generation DL models, and the speaker's background in developing performant and power-efficient deep learning processors. Gain insights into topics such as Moore's Law, obstacles in chip design, architectural solutions, new memory technologies, wireless links, thermal management, and comparisons with other approaches like TPU v1, TPU v2, and analog designs.

Syllabus

Intro
Deep Learning
A Specialized Chip
Ambitious Target
Moores Law
How
Obstacles
Criteria
Problem
Architecture
Highlevel overview
Circuit level improvements
New memory technology
Wireless link
Thermals
Thermal multiplexing
External memory
TPU v1
TPU v2
Analog approaches
Computational graph

Taught by

Stanford Online

Reviews

Start your review of Petascale Deep Learning on a Single Chip

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.