YoVDO

Efficient Stable Diffusion Pre-Training on Billions of Images with Ray

Offered By: Databricks via YouTube

Tags

Machine Learning Courses Stable Diffusion Courses Image Processing Courses Fault Tolerance Courses GPU Computing Courses Distributed Computing Courses Data Preprocessing Courses Heterogeneous Computing Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Discover how to efficiently pre-train Stable Diffusion models on billions of images using Ray in this 30-minute conference talk by Databricks. Learn to overcome challenges in scaling data preprocessing, improving GPU utilization, ensuring fault tolerance, and managing heterogeneous clusters. Explore an end-to-end pre-training solution that achieves large-scale state-of-the-art performance using Ray Data and Ray Train. Gain insights on implementing a stable diffusion pre-training pipeline, enhancing efficiency in large-scale multimodal data processing, and scaling online preprocessing and distributed training across different GPU types to optimize utilization and reduce costs. Presented by Hao Chen and Yunxuan Xiao from Anyscale Inc., this talk offers valuable takeaways for maximizing performance and cost efficiency in Stable Diffusion pre-training at scale.

Syllabus

Efficient Stable Diffusion Pre-Training on Billions of Images with Ray


Taught by

Databricks

Related Courses

Future of Computing - IBM Power 9 and beyond
openHPI
SIGCOMM 2020 - Reducto - On-Camera Filtering for Resource-Efficient Real-Time Video Analytics
Association for Computing Machinery (ACM) via YouTube
Offload Annotations - Bringing Heterogeneous Computing to Existing Libraries and Workloads
USENIX via YouTube
Supercomputing Spotlights - Supercomputing Software for Moore and Beyond
Society for Industrial and Applied Mathematics via YouTube
Liquid Metal - Taming Heterogeneity
GOTO Conferences via YouTube