Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Estimating Normalizing Constants for Log-Concave Distributions - Algorithms and Lower Bounds

Association for Computing Machinery (ACM) via YouTube

Overview

Explore algorithms and lower bounds for estimating normalizing constants of log-concave distributions in this 26-minute ACM conference talk. Delve into problem statement, upper bounds using annealing and multilevel Monte Carlo techniques, and regular Monte Carlo methods. Examine sampling algorithms, focusing on Langevin dynamics, its discretization, and coupling. Investigate lower bounds for both low and high dimensions, including proof ideas and distinguishing biased coins. Conclude with insights on partitioning dimensions and key takeaways for computational statistics and machine learning applications.

Syllabus

Intro
Problem statement
Upper bound: Annealing
Upper bound: Multilevel Monte Carlo
Regular Monte Carlo
Sampling algorithm: Langevin dynamics
Discretizing Langevin dynamics
Coupling Langevin dynamics (Overdamped)
Lower bound for low dimensions
Proof idea
Distinguishing biased coins
Lower bound for high dimensions Take product distribution Partition de dimensions into
Conclusion

Taught by

Association for Computing Machinery (ACM)

Reviews

Start your review of Estimating Normalizing Constants for Log-Concave Distributions - Algorithms and Lower Bounds

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.