Scalable Extraction of Training Data from Language Models
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore a detailed analysis of a research paper revealing how large language models like ChatGPT can inadvertently leak training data through simple prompts. Delve into the concepts of extractable and discoverable memorization, examining how researchers were able to extract gigabytes of data from various models. Learn about the novel "divergence attack" developed to exploit ChatGPT, causing it to deviate from normal behavior and emit training data at a much higher rate. Understand the implications of these findings for data privacy, model security, and the effectiveness of current alignment techniques in preventing memorization. Gain insights into quantitative membership testing and the broader consequences of this research for the field of AI and machine learning.
Syllabus
- Intro
- Extractable vs Discoverable Memorization
- Models leak more data than previously thought
- Some data is extractable but not discoverable
- Extracting data from closed models
- Poem poem poem
- Quantitative membership testing
- Exploring the ChatGPT exploit further
- Conclusion
Taught by
Yannic Kilcher
Related Courses
Microsoft Bot Framework and Conversation as a PlatformMicrosoft via edX Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube Improving Customer Experiences with Speech to Text and Text to Speech
Microsoft via YouTube Stanford Seminar - Deep Learning in Speech Recognition
Stanford University via YouTube Select Topics in Python: Natural Language Processing
Codio via Coursera