Defending Against Adversarial Model Attacks Using Kubeflow
Offered By: CNCF [Cloud Native Computing Foundation] via YouTube
Course Description
Overview
Explore a conference talk on defending against adversarial model attacks using Kubeflow. Learn about the importance of AI algorithm robustness in critical domains like self-driving cars, facial recognition, and hiring. Discover how to build a pipeline resistant to adversarial attacks by leveraging Kubeflow Pipelines and integrating with LFAI Adversarial Robustness Toolbox (ART). Gain insights into testing machine learning model's adversarial robustness in production on Kubeflow Serving using Payload logging and ART. Cover topics including Trusted AI, Open Governance, Security, Toolkit, and other related projects. Conclude with a Kubeflow survey and a practical demonstration.
Syllabus
Introduction
Trusted AI
Open Governance
Security
Toolkit
Other Projects
Adversarial robustness toolbox
Kubeflow Survey
Demo
Taught by
CNCF [Cloud Native Computing Foundation]
Related Courses
Computer SecurityStanford University via Coursera Cryptography II
Stanford University via Coursera Malicious Software and its Underground Economy: Two Sides to Every Story
University of London International Programmes via Coursera Building an Information Risk Management Toolkit
University of Washington via Coursera Introduction to Cybersecurity
National Cybersecurity Institute at Excelsior College via Canvas Network