Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion
Offered By: Valence Labs via YouTube
Course Description
Overview
Explore a comprehensive lecture on Diffusion Forcing, a novel training paradigm for sequence generative modeling. Delve into the concept of training a diffusion model to denoise tokens with independent per-token noise levels, combining the strengths of next-token prediction and full-sequence diffusion models. Learn about the method's ability to generate variable-length sequences, guide sampling towards desirable trajectories, and roll out continuous token sequences beyond the training horizon. Discover new sampling and guiding schemes unique to Diffusion Forcing's architecture, leading to improved performance in decision-making and planning tasks. Gain insights into the theoretical foundations of the approach, including its optimization of a variational lower bound on subsequence likelihoods. The lecture covers background information, the core principles of Diffusion Forcing, its application with causal uncertainty, and concludes with a Q&A session.
Syllabus
- Intro + Background
- Diffusion Forcing
- DF with Causal Uncertainty
- Q+A
Taught by
Valence Labs
Related Courses
Deep Learning – Part 2Indian Institute of Technology Madras via Swayam Probabilistic Deep Learning with TensorFlow 2
Imperial College London via Coursera Introduction to Deep Learning
Massachusetts Institute of Technology via YouTube Spatial Computational Thinking
National University of Singapore via edX DALL-E - Zero-Shot Text-to-Image Generation - Paper Explained
Aleksa Gordić - The AI Epiphany via YouTube