Mandoline - Model Evaluation under Distribution Shift
Offered By: Stanford University via YouTube
Course Description
Overview
Explore a comprehensive framework for evaluating machine learning models under distribution shift in this Stanford University lecture. Dive into the Mandoline approach, which leverages user-defined "slicing functions" to guide importance weighting and improve performance estimation on target distributions. Learn how this method outperforms standard baselines in NLP and vision tasks, and understand its connection to interactive ML systems. Gain insights into the theoretical foundations of the framework, including density ratio estimation and its error scaling. Discover the broader implications for model evaluation, hidden stratification, and iterative model development processes in the context of deploying ML models in real-world scenarios.
Syllabus
Intro
Outline
The ML model development process
Model Evaluation
Motivation
Common approach: importance weighting
Motivating example
Mandoline: Slice-based reweighting framework
The theory behind using slices
More formally...
Density Ratio Estimation
Experiments: tasks
Experiments: compare to reweighting on x
Summary
Taking a step back - how do we get slices? What are sli
Measuring model performance
Hidden Stratification: Approach
ML model development process, revisited
Another angle - how else can we evaluate?
"Closing the loop" - how do we update?
Taught by
Stanford MedAI
Tags
Related Courses
Natural Language ProcessingColumbia University via Coursera Natural Language Processing
Stanford University via Coursera Introduction to Natural Language Processing
University of Michigan via Coursera moocTLH: Nuevos retos en las tecnologĂas del lenguaje humano
Universidad de Alicante via MirĂadax Natural Language Processing
Indian Institute of Technology, Kharagpur via Swayam