YoVDO

Deep Learning Robustness Verification for Few-Pixel Attacks

Offered By: ACM SIGPLAN via YouTube

Tags

Neural Networks Courses Deep Learning Courses Dynamic programming Courses Adversarial Attacks Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a groundbreaking approach to verifying the robustness of neural networks against few-pixel attacks in this 18-minute video presentation from OOPSLA 2023. Delve into the innovative Calzone method, developed by researchers from Technion, Israel, which offers the first sound and complete analysis for L0 adversarial attacks. Learn how this technique leverages dynamic programming, sampling, and covering designs to efficiently verify network robustness, typically completing within minutes for most cases. Discover how Calzone outperforms existing methods, scaling to handle challenging instances where traditional approaches fail. Gain insights into the importance of robustness verification in deep learning and its implications for creating more secure and reliable neural networks.

Syllabus

[OOPSLA23] Deep Learning Robustness Verification for Few-Pixel Attacks


Taught by

ACM SIGPLAN

Related Courses

Machine Learning and Artificial Intelligence Security Risk: Categorizing Attacks and Failure Modes
LinkedIn Learning
How Apple Scans Your Phone and How to Evade It - NeuralHash CSAM Detection Algorithm Explained
Yannic Kilcher via YouTube
Deep Learning New Frontiers
Alexander Amini via YouTube
Deep Learning New Frontiers
Alexander Amini via YouTube
MIT 6.S191 - Deep Learning Limitations and New Frontiers
Alexander Amini via YouTube