Serverless Machine Learning Inference with KFServing

Serverless Machine Learning Inference with KFServing

CNCF [Cloud Native Computing Foundation] via YouTube Direct link

Challenge: Deploy many models

7 of 16

7 of 16

Challenge: Deploy many models

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Serverless Machine Learning Inference with KFServing

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Intro
  2. 2 Inference Stack Evolution PYTORCH
  3. 3 Model explanation, model pre-post transformers
  4. 4 GPU Autoscaling the challenge
  5. 5 Challenge: Increase GPU utilization
  6. 6 Use Case: Personalized News Monitoring
  7. 7 Challenge: Deploy many models
  8. 8 Proposed Solution: Multi-model Inference Service
  9. 9 Experience from running a serverless inference platform
  10. 10 Reduce tail latency caused by CPU throttling
  11. 11 Reduce cold start latency
  12. 12 Monitoring and Alerting: Control Plane
  13. 13 Monitoring and Alerting: Access logs
  14. 14 Monitoring and Alerting: Inference Service metrics
  15. 15 KFServing Roadmap 2020
  16. 16 Our Working Group is Open

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.