YoVDO

Evaluating NLP Models via Contrast Sets

Offered By: Yannic Kilcher via YouTube

Tags

Model Evaluation Courses Linguistic Analysis Courses

Course Description

Overview

Explore a 19-minute video lecture on evaluating natural language processing (NLP) models using contrast sets. Learn about the limitations of current evaluation methods for supervised learning tasks in NLP and how models often exploit dataset-specific correlations rather than learning the intended task. Discover the concept of contrast sets - hand-crafted perturbations created by dataset authors to capture their original intent and provide a more meaningful evaluation of model performance. Examine the proposed annotation paradigm for creating contrast sets and its application to 10 diverse NLP datasets. Understand how contrast sets offer a local view of a model's decision boundary and can reveal significant drops in performance compared to standard test sets. Gain insights into improving NLP model evaluation and dataset construction to better assess true linguistic capabilities.

Syllabus

Evaluating NLP Models via Contrast Sets


Taught by

Yannic Kilcher

Related Courses

Macroeconometric Forecasting
International Monetary Fund via edX
Machine Learning With Big Data
University of California, San Diego via Coursera
Data Science at Scale - Capstone Project
University of Washington via Coursera
Structural Equation Model and its Applications | 结构方程模型及其应用 (粤语)
The Chinese University of Hong Kong via Coursera
Data Science in Action - Building a Predictive Churn Model
SAP Learning