HyperTransformer - Model Generation for Supervised and Semi-Supervised Few-Shot Learning

HyperTransformer - Model Generation for Supervised and Semi-Supervised Few-Shot Learning

Yannic Kilcher via YouTube Direct link

- How could we produce larger weights?

10 of 12

10 of 12

- How could we produce larger weights?

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

HyperTransformer - Model Generation for Supervised and Semi-Supervised Few-Shot Learning

Automatically move to the next video in the Classroom when playback concludes

  1. 1 - Intro & Overview
  2. 2 - Weight-generation vs Fine-tuning for few-shot learning
  3. 3 - HyperTransformer model architecture overview
  4. 4 - Why the self-attention mechanism is useful here
  5. 5 - Start of Interview
  6. 6 - Can neural networks even produce weights of other networks?
  7. 7 - How complex does the computational graph get?
  8. 8 - Why are transformers particularly good here?
  9. 9 - What can the attention maps tell us about the algorithm?
  10. 10 - How could we produce larger weights?
  11. 11 - Diving into experimental results
  12. 12 - What questions remain open?

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.