YoVDO

Approximate Caching for Efficiently Serving Text-to-Image Diffusion Models

Offered By: USENIX via YouTube

Tags

Machine Learning Courses Computer Vision Courses Image Processing Courses Diffusion Models Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a conference talk presenting NIRVANA, an innovative system for text-to-image generation using diffusion models. Discover how approximate caching techniques can significantly reduce computational requirements and latency in producing high-quality images from text prompts. Learn about the novel cache management policy that enables 21% GPU compute savings, 19.8% end-to-end latency reduction, and 19% cost savings in real production environments. Gain insights into the extensive characterization of production text-to-image prompts, focusing on caching, popularity, and reuse of intermediate states in large-scale deployments. Understand the potential impact of this research on making resource-intensive diffusion models more efficient and accessible for various applications.

Syllabus

NSDI '24 - Approximate Caching for Efficiently Serving Text-to-Image Diffusion Models


Taught by

USENIX

Related Courses

Introduction to Artificial Intelligence
Stanford University via Udacity
Computer Vision: The Fundamentals
University of California, Berkeley via Coursera
Computational Photography
Georgia Institute of Technology via Coursera
Digital Signal Processing
École Polytechnique Fédérale de Lausanne via Coursera
Creative, Serious and Playful Science of Android Apps
University of Illinois at Urbana-Champaign via Coursera