YoVDO

Data Augmentation MCMC for Bayesian Inference from Privatized Data

Offered By: Fields Institute via YouTube

Tags

Bayesian Inference Courses Data Privacy Courses Differential Privacy Courses Naive Bayes Classifier Courses

Course Description

Overview

Explore a 28-minute conference talk by Ruobin Gong from Rutgers University on Data Augmentation MCMC for Bayesian Inference from Privatized Data. Delve into the challenges of modern data curation, focusing on differential privacy and its adoption by the U.S. Census Bureau. Examine the mechanism, benefits, and challenges of differential privacy, and learn about statistical inference from privatized data. Discover existing solutions, including traditional Gibbs sampling and a general Metropolis-within Gibbs sampler. Analyze the requirements, run time, and efficiency of these methods, as well as the ergodicity of the proposed sampler. Apply the concepts to a naïve Bayes classifier through a simulation setup, exploring posterior mean, frequentist coverage, and empirical acceptance rates. Gain valuable insights into the intersection of statistical analysis and privacy in data science.

Syllabus

Intro
Privacy: a challenge in modern data curation
The U.S. Census Bureau adopts differential privacy
The mechanism of differential privacy
Differential privacy: benefits and challenges
Situating our (statistical x privacy) framework
Statistical inference from privatized data
Existing solutions
A traditional Gibbs sampler
A general Metropolis-within Gibbs sampler
Requirements, run time, and efficiency
Ergodicity of the proposed sampler
Application: a naïve Bayes classifier
Simulation setup
Posterior mean
Frequentist coverage
Empirical acceptance rates
Summary


Taught by

Fields Institute

Related Courses

Statistical Machine Learning
Carnegie Mellon University via Independent
Secure and Private AI
Facebook via Udacity
Data Privacy and Anonymization in R
DataCamp
Build and operate machine learning solutions with Azure Machine Learning
Microsoft via Microsoft Learn
Data Privacy and Anonymization in Python
DataCamp