YoVDO

Fairness Through Robustness - Investigating Robustness Disparity in Deep Learning

Offered By: Association for Computing Machinery (ACM) via YouTube

Tags

ACM FAccT Conference Courses Data Analysis Courses Research Methodology Courses Fairness in AI Courses

Course Description

Overview

Explore a 21-minute conference talk that delves into the intersection of fairness and robustness in deep learning systems. Investigate how robustness disparity affects different demographic groups and its implications for algorithmic fairness. Learn about data preliminaries, metrics, and key research questions addressed in this study. Examine the results obtained from experiments on various datasets, including the intriguing "Aliens" data. Discover potential mitigation strategies to address robustness disparities and enhance fairness in machine learning models. Gain insights into the critical relationship between robustness and fairness in AI systems, presented by researchers V. Nanda, S. Dooley, S. Singla, S. Feizi, and J. Dickerson at the FAccT 2021 virtual conference.

Syllabus

Introduction
Data Preliminaries
Metrics
Research Questions
Results
Aliens Data
Mitigation


Taught by

ACM FAccT Conference

Related Courses

Translation Tutorial - Thinking Through and Writing About Research Ethics Beyond "Broader Impact"
Association for Computing Machinery (ACM) via YouTube
Translation Tutorial - Data Externalities
Association for Computing Machinery (ACM) via YouTube
Translation Tutorial - Causal Fairness Analysis
Association for Computing Machinery (ACM) via YouTube
Implications Tutorial - Using Harms and Benefits to Ground Practical AI Fairness Assessments
Association for Computing Machinery (ACM) via YouTube
Responsible AI in Industry - Lessons Learned in Practice
Association for Computing Machinery (ACM) via YouTube