Mandoline - Model Evaluation under Distribution Shift
Offered By: Stanford University via YouTube
Course Description
Overview
Explore a comprehensive framework for evaluating machine learning models under distribution shift in this Stanford University lecture. Dive into the Mandoline approach, which leverages user-defined "slicing functions" to guide importance weighting and improve performance estimation on target distributions. Learn how this method outperforms standard baselines in NLP and vision tasks, and understand its connection to interactive ML systems. Gain insights into the theoretical foundations of the framework, including density ratio estimation and its error scaling. Discover the broader implications for model evaluation, hidden stratification, and iterative model development processes in the context of deploying ML models in real-world scenarios.
Syllabus
Intro
Outline
The ML model development process
Model Evaluation
Motivation
Common approach: importance weighting
Motivating example
Mandoline: Slice-based reweighting framework
The theory behind using slices
More formally...
Density Ratio Estimation
Experiments: tasks
Experiments: compare to reweighting on x
Summary
Taking a step back - how do we get slices? What are sli
Measuring model performance
Hidden Stratification: Approach
ML model development process, revisited
Another angle - how else can we evaluate?
"Closing the loop" - how do we update?
Taught by
Stanford MedAI
Tags
Related Courses
Macroeconometric ForecastingInternational Monetary Fund via edX Machine Learning With Big Data
University of California, San Diego via Coursera Data Science at Scale - Capstone Project
University of Washington via Coursera Structural Equation Model and its Applications | 结构方程模型及其应用 (粤语)
The Chinese University of Hong Kong via Coursera Data Science in Action - Building a Predictive Churn Model
SAP Learning