Approximate Caching for Efficiently Serving Text-to-Image Diffusion Models
Offered By: USENIX via YouTube
Course Description
Overview
Explore a conference talk presenting NIRVANA, an innovative system for text-to-image generation using diffusion models. Discover how approximate caching techniques can significantly reduce computational requirements and latency in producing high-quality images from text prompts. Learn about the novel cache management policy that enables 21% GPU compute savings, 19.8% end-to-end latency reduction, and 19% cost savings in real production environments. Gain insights into the extensive characterization of production text-to-image prompts, focusing on caching, popularity, and reuse of intermediate states in large-scale deployments. Understand the potential impact of this research on making resource-intensive diffusion models more efficient and accessible for various applications.
Syllabus
NSDI '24 - Approximate Caching for Efficiently Serving Text-to-Image Diffusion Models
Taught by
USENIX
Related Courses
Diffusion Models Beat GANs on Image Synthesis - Machine Learning Research Paper ExplainedYannic Kilcher via YouTube Diffusion Models Beat GANs on Image Synthesis - ML Coding Series - Part 2
Aleksa Gordić - The AI Epiphany via YouTube OpenAI GLIDE - Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models
Aleksa Gordić - The AI Epiphany via YouTube Food for Diffusion
HuggingFace via YouTube Imagen: Text-to-Image Generation Using Diffusion Models - Lecture 9
University of Central Florida via YouTube