Scaling Training and Batch Inference - A Deep Dive into Ray AIR's Data Processing Engine
Offered By: Anyscale via YouTube
Course Description
Overview
Dive into a comprehensive 44-minute talk exploring AIR's data processing engine for scaling training and batch inference. Learn how Ray AIR leverages Ray Datasets to achieve high performance and scalability in ML pipelines. Discover techniques for efficient data loading and preprocessing across multiple machines, addressing ingest bottlenecks and maximizing GPU utilization. Explore key features like distributed data sharding, parallel I/O, transformations, CPU-GPU compute pipelining, autoscaling inference workers, and efficient per-epoch shuffling. Gain insights from real-world case studies of production AIR workloads, showcasing performance and scalability benefits. Master the creation of scalable training and batch inference pipelines using Ray AIR to optimize your machine learning workflows.
Syllabus
Intro
Overview
ML Pipelines Must Scale with Data
Distributed Data-Parallel to the Rescue
Scaling the Typical ML Pipeline
Possible Solution - Coordinated Pipelining
Ray Datasets: AIR's Data Processing Engine
Avoiding GPU Data Prep Stalls
Dataset Sharding
Parallel I/O and Transformations
Dataplane Optimizations
Pipelining Ingest with Training
Pipelining Ingest with Inference
Autoscaling Actor Pool for Inference
Per-epoch Shuffling - Distributed
ML engineer at Telematics Startup
Summary
Taught by
Anyscale
Related Courses
Cloud Computing Concepts, Part 1University of Illinois at Urbana-Champaign via Coursera Cloud Computing Concepts: Part 2
University of Illinois at Urbana-Champaign via Coursera Reliable Distributed Algorithms - Part 1
KTH Royal Institute of Technology via edX Introduction to Apache Spark and AWS
University of London International Programmes via Coursera Réalisez des calculs distribués sur des données massives
CentraleSupélec via OpenClassrooms