Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Recent Progress in High-Dimensional Learning

Simons Institute via YouTube

Overview

Explore recent advancements in high-dimensional learning through this comprehensive lecture by MIT's Ankur Moitra. Delve into parametric statistics, maximum likelihood estimation, and noise constraint techniques. Examine the effectiveness of empirical mean and variance, and understand the Folklore Theorem. Investigate robustness, hardness, and the price of robustness in estimation. Learn about the Robust Estimation Recipe and the WinWin Algorithm. Analyze relaxed distributional assumptions and their impact on robust estimation. Consider the error guarantee tendencies and the completeness of adversaries in gaussian mean estimation. Gain insights into adversary efficiency and improper learning in high-dimensional contexts.

Syllabus

Introduction
Parametric Statistics
Maximum likelihood estimation
How to constrain noise
Estimating parameters
Do empirical mean and empirical variance work
Folklore Theorem
Robustness and Hardness
Price of Robustness
Recent Results
Robust Estimation Recipe
WinWin Algorithm
Birds Eye View
O of epsilon
Relaxing Distributional Assumption
Robust Estimation
Conclusion
Does the error guarantee tend to O
Is there a sense that some adversaries are complete
gaussian mean estimation
adversary efficiency
improper learning

Taught by

Simons Institute

Reviews

Start your review of Recent Progress in High-Dimensional Learning

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.