Privately Evaluating Untrusted Black-Box Functions
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore a plenary talk on privately evaluating untrusted black-box functions in the context of differential privacy. Delve into tools for sharing sensitive data when data curators face unknown queries from untrusted analysts. Learn about privacy wrappers, algorithms that provide differentially private approximations of black-box functions on sensitive datasets. Examine two settings: automated sensitivity detection and provided sensitivity bounds. Discover the first privacy wrapper for automated sensitivity detection and improvements in accuracy and query complexity for the provided sensitivity bound setting. Investigate lower bounds for both scenarios and understand how these mechanisms contribute to the feasibility of differentially private release of general function classes.
Syllabus
Plenary Talk: Privately Evaluating Untrusted Black-Box Functions
Taught by
Simons Institute
Related Courses
Statistical Machine LearningCarnegie Mellon University via Independent Secure and Private AI
Facebook via Udacity Data Privacy and Anonymization in R
DataCamp Build and operate machine learning solutions with Azure Machine Learning
Microsoft via Microsoft Learn Data Privacy and Anonymization in Python
DataCamp