YoVDO

Stop Explaining Black Box ML Models for High Stakes Decisions and Use Interpretable Models

Offered By: Toronto Machine Learning Series (TMLS) via YouTube

Tags

Explainable AI Courses Machine Learning Courses Criminal Justice Courses Ethics in AI Courses Algorithmic Fairness Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the critical implications of using black box machine learning models for high-stakes decisions in this thought-provoking 49-minute conference talk from the Toronto Machine Learning Series. Delve into the insights of Cynthia Rudin, Professor of Computer Science, Electrical and Computer Engineering, and Statistical Science at Duke University, as she challenges the widespread use of opaque ML models. Examine the serious societal consequences, including flawed bail and parole decisions in criminal justice, that arise from relying on these models. Discover why explanations for black box models can be unreliable and potentially misleading. Learn about the advantages of interpretable machine learning models, which provide inherent explanations faithful to their actual computations. Gain valuable perspectives on the importance of transparency and accountability in AI-driven decision-making processes for high-stakes scenarios.

Syllabus

Stop Explaining Black Box ML Models for High Stakes Decisions and Use Interpretable Models


Taught by

Toronto Machine Learning Series (TMLS)

Related Courses

Artificial Intelligence Privacy and Convenience
LearnQuest via Coursera
Capstone Assignment - CDSS 5
University of Glasgow via Coursera
Learn Explainable AI (XAI)
Codecademy
Responsible AI for Developers: Interpretability & Transparency - 简体中文
Google Cloud via Coursera
Developing Explainable AI (XAI)
Duke University via Coursera