Private Convex Optimization via Exponential Mechanism - Differential Privacy for Machine Learning
Offered By: Google TechTalks via YouTube
Course Description
Overview
Explore private convex optimization through the exponential mechanism in this Google TechTalk presented by Daogao Liu. Delve into differential privacy for machine learning, covering topics such as noisy stochastic gradient descent and the regularized exponential mechanism. Examine isoperimetric inequality for strongly log-concave measures and concentration bounds for Lipschitz functions. Learn about DP-Stochastic Convex Optimization and its intuition. Discover new sampling algorithms and their applications in DP-ERM and DP-SCO. Gain insights into bounding generalization error, Wasserstein distance, KL divergence, and population loss. Understand the contributions and open problems in this field of private convex optimization.
Syllabus
Intro
One-sentence Summary
Differential Privacy
Noisy SGD
Regularized Exponential Mechanism (RegEM)
Isoperimetric Inequality for strongly log- concave measures
Concentration bounds for Lipschitz functions
Proof Sketch
Utility Analysis
A Question from the Duck
DP-Stochastic Convex Optimization (SCO)
Intuition
Open Problems
RegEM Revisited
Bounding Generalization Error
Bound Wasserstein Distance
Bounding KL. divergence
Bounding Population Loss
Summary of Contributions
A new sampling algorithm
Algorithms for DP-ERM and DP-SCO
Taught by
Google TechTalks
Related Courses
Building Classification Models with scikit-learnPluralsight Practical Deep Learning for Coders - Full Course
freeCodeCamp Neural Networks Made Easy
Udemy Intro to Deep Learning
Kaggle Stochastic Gradient Descent
Great Learning via YouTube