YoVDO

Estimating Normalizing Constants for Log-Concave Distributions - Algorithms and Lower Bounds

Offered By: Association for Computing Machinery (ACM) via YouTube

Tags

Machine Learning Courses Algorithms Courses Langevin Dynamics Courses

Course Description

Overview

Explore algorithms and lower bounds for estimating normalizing constants of log-concave distributions in this 26-minute ACM conference talk. Delve into problem statement, upper bounds using annealing and multilevel Monte Carlo techniques, and regular Monte Carlo methods. Examine sampling algorithms, focusing on Langevin dynamics, its discretization, and coupling. Investigate lower bounds for both low and high dimensions, including proof ideas and distinguishing biased coins. Conclude with insights on partitioning dimensions and key takeaways for computational statistics and machine learning applications.

Syllabus

Intro
Problem statement
Upper bound: Annealing
Upper bound: Multilevel Monte Carlo
Regular Monte Carlo
Sampling algorithm: Langevin dynamics
Discretizing Langevin dynamics
Coupling Langevin dynamics (Overdamped)
Lower bound for low dimensions
Proof idea
Distinguishing biased coins
Lower bound for high dimensions Take product distribution Partition de dimensions into
Conclusion


Taught by

Association for Computing Machinery (ACM)

Related Courses

Information Theory
The Chinese University of Hong Kong via Coursera
Intro to Computer Science
University of Virginia via Udacity
Analytic Combinatorics, Part I
Princeton University via Coursera
Algorithms, Part I
Princeton University via Coursera
Divide and Conquer, Sorting and Searching, and Randomized Algorithms
Stanford University via Coursera