Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Machine Learning: Computational Learning Theory and Batch Learning - Lecture 12

UofU Data Science via YouTube

Overview

Explore formal models of learnability in this 48-minute machine learning lecture focusing on batch learning and computational learning theory. Delve into the mistake bound model before examining what formal models of learnability provide and understanding the goals of batch learning. Learn about learning conjunctions, random source training examples, nature-provided labels, and different approaches to evaluating learning algorithms. Understand key concepts including distribution over instance space, PAC Learning intuition, hypothesis error, empirical error, and the distinctions between online and batch learning methodologies.

Syllabus

Intro
Checkpoint: The bigger picture
Learning Conjunctions Some random source (nature) provides training examples Teacher (Nature) provides the labels (f(x))
Two Directions for How good is our learning algorithm?
The mistake bound approach
The setup
Distribution over the instance space
PAC Learning - Intuition
Error of a hypothesis
Empirical error
The goal of batch learning
Online learning vs. Batch learning

Taught by

UofU Data Science

Reviews

Start your review of Machine Learning: Computational Learning Theory and Batch Learning - Lecture 12

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.