YoVDO

Approximate Caching for Efficiently Serving Text-to-Image Diffusion Models

Offered By: USENIX via YouTube

Tags

Machine Learning Courses Computer Vision Courses Image Processing Courses Diffusion Models Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a conference talk presenting NIRVANA, an innovative system for text-to-image generation using diffusion models. Discover how approximate caching techniques can significantly reduce computational requirements and latency in producing high-quality images from text prompts. Learn about the novel cache management policy that enables 21% GPU compute savings, 19.8% end-to-end latency reduction, and 19% cost savings in real production environments. Gain insights into the extensive characterization of production text-to-image prompts, focusing on caching, popularity, and reuse of intermediate states in large-scale deployments. Understand the potential impact of this research on making resource-intensive diffusion models more efficient and accessible for various applications.

Syllabus

NSDI '24 - Approximate Caching for Efficiently Serving Text-to-Image Diffusion Models


Taught by

USENIX

Related Courses

Introduction to Artificial Intelligence
Stanford University via Udacity
Natural Language Processing
Columbia University via Coursera
Probabilistic Graphical Models 1: Representation
Stanford University via Coursera
Computer Vision: The Fundamentals
University of California, Berkeley via Coursera
Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent