Gradient Obfuscation Gives a False Sense of Security in Federated Learning
Offered By: USENIX via YouTube
Course Description
Overview
Explore a critical analysis of privacy protection mechanisms in federated learning presented at USENIX Security '23. Delve into a new reconstruction attack framework for image classification tasks that challenges the effectiveness of gradient obfuscation techniques. Examine how common gradient postprocessing procedures, including quantization, sparsification, and perturbation, may provide a false sense of security. Discover a novel method for reconstructing images at the semantic level and learn about the quantification of semantic privacy leakage. Compare this approach with conventional image similarity scores and understand the implications for evaluating image data leakage in federated learning. Gain insights into the urgent need for revisiting and redesigning privacy protection mechanisms in existing federated learning algorithms to ensure robust client data security.
Syllabus
USENIX Security '23 - Gradient Obfuscation Gives a False Sense of Security in Federated Learning
Taught by
USENIX
Related Courses
Secure and Private AIFacebook via Udacity Advanced Deployment Scenarios with TensorFlow
DeepLearning.AI via Coursera Big Data for Reliability and Security
Purdue University via edX MLOps for Scaling TinyML
Harvard University via edX Edge Analytics: IoT and Data Science
LinkedIn Learning