Robust Distortion-free Watermarks for Language Models
Offered By: Google TechTalks via YouTube
Course Description
Overview
Explore a Google TechTalk presented by John Thickstun on robust distortion-free watermarks for language models. Delve into a protocol for planting watermarks into text generated by autoregressive language models that remain robust to edits without altering the distribution of generated text. Learn how the watermarking process controls the source of randomness using a secret key during the language model's decoding phase. Discover the statistical correlations used for watermark detection and the provable undetectability for those without the key. Examine two alternative decoders: inverse transform sampling and Gumbel argmax sampling. Gain insights from experimental validations using OPT-1.3B, LLaMA 7B, and Alpaca 7B language models, demonstrating statistical power and robustness against paraphrasing attacks. Understand the speaker's background as a postdoctoral researcher at Stanford University, his previous work, and recognition in the field of generative models and controllability.
Syllabus
Robust Distortion-free Watermarks for Language Models
Taught by
Google TechTalks
Related Courses
Introduction to Operations ManagementWharton School of the University of Pennsylvania via Coursera Computational Molecular Evolution
Technical University of Denmark (DTU) via Coursera Structural Equation Model and its Applications | 结构方程模型及其应用 (普通话)
The Chinese University of Hong Kong via Coursera Fundamentals of Clinical Trials
Harvard University via edX Curso Práctico de Bioestadística con R
Universidad San Pablo CEU via Miríadax