Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

How Far Can We Scale Up? Deep Learning's Diminishing Returns

Yannic Kilcher via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the limits of exponential scaling in AI and potential solutions in this 20-minute video review of an article on deep learning's diminishing returns. Examine the impressive results achieved through massive increases in computational power and data, while considering the challenges of overparameterization, power usage, and CO2 emissions. Delve into current attempts to address scaling issues, including a discussion on ImageNet V2 and the potential of symbolic methods. Gain insights into the future of AI development and the need for more efficient approaches to continue advancing the field.

Syllabus

- Intro & Overview
- Deep Learning at its limits
- The cost of overparameterization
- Extrapolating power usage and CO2 emissions
- We cannot just continue scaling up
- Current solution attempts
- Aside: ImageNet V2
- Are symbolic methods the way out?

Taught by

Yannic Kilcher

Reviews

Start your review of How Far Can We Scale Up? Deep Learning's Diminishing Returns

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.