Efficient Stable Diffusion Pre-Training on Billions of Images with Ray
Offered By: Databricks via YouTube
Course Description
Overview
Discover how to efficiently pre-train Stable Diffusion models on billions of images using Ray in this 30-minute conference talk by Databricks. Learn to overcome challenges in scaling data preprocessing, improving GPU utilization, ensuring fault tolerance, and managing heterogeneous clusters. Explore an end-to-end pre-training solution that achieves large-scale state-of-the-art performance using Ray Data and Ray Train. Gain insights on implementing a stable diffusion pre-training pipeline, enhancing efficiency in large-scale multimodal data processing, and scaling online preprocessing and distributed training across different GPU types to optimize utilization and reduce costs. Presented by Hao Chen and Yunxuan Xiao from Anyscale Inc., this talk offers valuable takeaways for maximizing performance and cost efficiency in Stable Diffusion pre-training at scale.
Syllabus
Efficient Stable Diffusion Pre-Training on Billions of Images with Ray
Taught by
Databricks
Related Courses
Future of Computing - IBM Power 9 and beyondopenHPI SIGCOMM 2020 - Reducto - On-Camera Filtering for Resource-Efficient Real-Time Video Analytics
Association for Computing Machinery (ACM) via YouTube Offload Annotations - Bringing Heterogeneous Computing to Existing Libraries and Workloads
USENIX via YouTube Supercomputing Spotlights - Supercomputing Software for Moore and Beyond
Society for Industrial and Applied Mathematics via YouTube Liquid Metal - Taming Heterogeneity
GOTO Conferences via YouTube