Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Automated Machine Learning Performance Evaluation

CNCF [Cloud Native Computing Foundation] via YouTube

Overview

Explore automated machine learning performance evaluation in this 26-minute conference talk from KubeCon + CloudNativeCon North America 2021. Dive into the intricacies of benchmarking deployed production machine learning models in cloud native infrastructure. Learn about the theory behind ML model benchmarking, including key parameters like latency, throughput, and performance percentiles. Follow a hands-on example using Argo, Kubernetes, and Seldon Core to benchmark a model across multiple parameters for optimal hardware performance. Gain insights into workflow management, reusability, and best practices for evaluating ML models in various deployment scenarios.

Syllabus

Introduction
C410 classifier
What is seldomcore
Deploying a model
Extra complexity
Best practices
Benchmark types
Benchmark tools
Automating the evaluation
Workflow managers
Workflows
argo workflow
reusability
output
resources
Wrap up

Taught by

CNCF [Cloud Native Computing Foundation]

Reviews

Start your review of Automated Machine Learning Performance Evaluation

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.