Discover how to evaluate enterprise Large Language Models (LLMs) using Snorkel Flow in this 20-minute demonstration video. Follow along as Snorkel AI software engineer Rebecca Westerlind guides you through the iterative loop at the core of Snorkel Flow's AI data development workflow. Learn to build a robust evaluation framework, leverage Snorkel's features for creating high-quality training data, and analyze LLM performance across various metrics and data slices. Gain practical insights into LLM evaluation, Snorkel Flow usage, and enterprise LLM deployment. This demo, an excerpt from a longer webinar, provides a step-by-step process to help you bridge the gap between demonstration and production-ready enterprise LLM applications.
Overview
Syllabus
DEMO: How to Evaluate Enterprise LLMs in Snorkel Flow
Taught by
Snorkel AI