YoVDO

Scalable Extraction of Training Data from Language Models

Offered By: Yannic Kilcher via YouTube

Tags

Language Models Courses Machine Learning Courses ChatGPT Courses Data Extraction Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a detailed analysis of a research paper revealing how large language models like ChatGPT can inadvertently leak training data through simple prompts. Delve into the concepts of extractable and discoverable memorization, examining how researchers were able to extract gigabytes of data from various models. Learn about the novel "divergence attack" developed to exploit ChatGPT, causing it to deviate from normal behavior and emit training data at a much higher rate. Understand the implications of these findings for data privacy, model security, and the effectiveness of current alignment techniques in preventing memorization. Gain insights into quantitative membership testing and the broader consequences of this research for the field of AI and machine learning.

Syllabus

- Intro
- Extractable vs Discoverable Memorization
- Models leak more data than previously thought
- Some data is extractable but not discoverable
- Extracting data from closed models
- Poem poem poem
- Quantitative membership testing
- Exploring the ChatGPT exploit further
- Conclusion


Taught by

Yannic Kilcher

Related Courses

Data Wrangling with MongoDB
MongoDB via Udacity
Data Science Essentials for SAP
OnSAP Academy via Independent
Herramientas de la Inteligencia de Negocios
Galileo University via edX
Digital Media Analytics: Using 'Listening Data'
Purdue University via FutureLearn
Advanced Business Analytics
University of Colorado Boulder via Coursera