YoVDO

Robustness to Adversarial Inputs and Tail Risk via Boosting - Machine Learning Model Defense Strategies

Offered By: Open Data Science via YouTube

Tags

Machine Learning Courses Neural Networks Courses Ensemble Methods Courses Class Imbalances Courses Algorithmic Fairness Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a 47-minute conference talk by Dr. Pradeep Ravikumar on enhancing machine learning model robustness against adversarial inputs and tail risk. Delve into innovative strategies for deploying ML models in high-stakes environments, focusing on an ensemble approach using neural networks to defend against threats while maintaining performance on challenging samples. Gain insights into critical areas such as algorithmic fairness, class imbalance, and risk-sensitive decision-making. Cover topics including distribution shift, unknown sub-populations, CVAR&DRO, randomized approaches, solving zero-sum games, and Distributional Outlier Robust Optimization (DORO). Ideal for professionals and enthusiasts in machine learning, AI, data science, and related fields seeking to advance their understanding of robust AI techniques and adversarial defense strategies.

Syllabus

- Intro
- Distribution Shift
- Unknown Sub-populations, Tail Risk
- CVAR&DRO
- Deterministic to Randomized
- Solving Zero Sum Games
- DORO: DIstributional Outlier Robust Optimization
- Summary


Taught by

Open Data Science

Related Courses

Neural Networks for Machine Learning
University of Toronto via Coursera
Good Brain, Bad Brain: Basics
University of Birmingham via FutureLearn
Statistical Learning with R
Stanford University via edX
Machine Learning 1—Supervised Learning
Brown University via Udacity
Fundamentals of Neuroscience, Part 2: Neurons and Networks
Harvard University via edX