Private Convex Optimization via Exponential Mechanism - Differential Privacy for Machine Learning
Offered By: Google TechTalks via YouTube
Course Description
Overview
Explore private convex optimization through the exponential mechanism in this Google TechTalk presented by Daogao Liu. Delve into differential privacy for machine learning, covering topics such as noisy stochastic gradient descent and the regularized exponential mechanism. Examine isoperimetric inequality for strongly log-concave measures and concentration bounds for Lipschitz functions. Learn about DP-Stochastic Convex Optimization and its intuition. Discover new sampling algorithms and their applications in DP-ERM and DP-SCO. Gain insights into bounding generalization error, Wasserstein distance, KL divergence, and population loss. Understand the contributions and open problems in this field of private convex optimization.
Syllabus
Intro
One-sentence Summary
Differential Privacy
Noisy SGD
Regularized Exponential Mechanism (RegEM)
Isoperimetric Inequality for strongly log- concave measures
Concentration bounds for Lipschitz functions
Proof Sketch
Utility Analysis
A Question from the Duck
DP-Stochastic Convex Optimization (SCO)
Intuition
Open Problems
RegEM Revisited
Bounding Generalization Error
Bound Wasserstein Distance
Bounding KL. divergence
Bounding Population Loss
Summary of Contributions
A new sampling algorithm
Algorithms for DP-ERM and DP-SCO
Taught by
Google TechTalks
Related Courses
Information TheoryNPTEL via Swayam Reinforcement Learning from Human Feedback - From Zero to ChatGPT
HuggingFace via YouTube What Is Quantum Machine Learning?
Serrano.Academy via YouTube Robust Pre-Training by Adversarial Contrastive Learning - CAP6412 Spring 2021
University of Central Florida via YouTube Variational Autoencoders
Paul Hand via YouTube