Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks
Offered By: IEEE via YouTube
Course Description
Overview
Explore a 22-minute IEEE conference talk on defensive distillation as a countermeasure against adversarial attacks on deep neural networks. Delve into the vulnerability of deep learning algorithms to adversarial samples and their potential security implications. Learn about the proposed defensive distillation mechanism and its effectiveness in reducing the impact of adversarial inputs. Examine analytical and empirical studies demonstrating how this technique can dramatically decrease the success rate of sample creation from 95% to less than 0.5% and increase the complexity of creating adversarial samples. Gain insights into the generalizability and robustness properties granted by defensive distillation in training deep neural networks, and understand its potential applications in enhancing the security of AI-powered systems.
Syllabus
Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks
Taught by
IEEE Symposium on Security and Privacy
Tags
Related Courses
Computer SecurityStanford University via Coursera Cryptography II
Stanford University via Coursera Malicious Software and its Underground Economy: Two Sides to Every Story
University of London International Programmes via Coursera Building an Information Risk Management Toolkit
University of Washington via Coursera Introduction to Cybersecurity
National Cybersecurity Institute at Excelsior College via Canvas Network