Machine Learning and Artificial Intelligence Security Risk: Categorizing Attacks and Failure Modes
Offered By: LinkedIn Learning
Course Description
Overview
Learn how and why machine learning and artificial intelligence technology fails and understand ways to make these systems more secure and resilient.
Syllabus
Introduction
- Machine learning security concerns
- What you should know
- How systems can fail and how to protect them
- Why does ML security matter
- Attacks vs. unintentional failure modes
- Security goals for ML: CIA
- Perturbation attacks and AUPs
- Poisoning attacks
- Reprogramming neural nets
- Physical domain (3D adversarial objects)
- Supply chain attacks
- Model inversion
- System manipulation
- Membership inference and model stealing
- Backdoors and existing exploits
- Reward hacking
- Side effects in reinforcement learning
- Distributional shifts and incomplete testing
- Overfitting/underfitting
- Data bias considerations
- Effective techniques for building resilience in ML
- ML dataset hygiene
- ML adversarial training
- ML access control to APIs
- Next steps
Taught by
Diana Kelley
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Natural Language Processing
Columbia University via Coursera Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent