Scaling Python for Machine Learning - Beyond Data Parallelism
Offered By: GOTO Conferences via YouTube
Course Description
Overview
Explore advanced techniques for scaling Python in machine learning beyond data parallelism in this conference talk from GOTO Chicago 2023. Dive into the world of distributed computing with Holden Karau, an Open Source Engineer at Netflix, as she examines Spark, Dask, and Ray for scaling machine learning models. Learn about distributed tasks, actor models for managing model weights during training, and fault tolerance in various frameworks. Gain insights into the similarities and differences between Dask and Ray distributed tasks, understand task and actor fault tolerance, and discover the relationship between Ray and Netflix. This comprehensive presentation covers topics such as data parallelism refresher, distributed task structures, Spark's capabilities, and actor fault tolerance in Ray and Dask, providing valuable knowledge for scaling Python applications in machine learning contexts.
Syllabus
Intro
Probable relevant biases
Quick refresher on data parallelism
What do distributed tasks look like?
Dask distributed tasks
Ray distributed tasks
How are they different & same?
Task fault tolerance
Does Spark have tasks & actors?
Ray Diagram
Ray actor fault tolerance
What's up with Ray & Netflix?
Dask actor fault tolerance
Outro
Taught by
GOTO Conferences
Related Courses
MongoDB for DBAsMongoDB University MongoDB Advanced Deployment and Operations
MongoDB University Building Cloud Apps with Microsoft Azure - Part 3
Microsoft via edX Implementing Microsoft Windows Server Disks and Volumes
Microsoft via edX Cloud Computing and Distributed Systems
Indian Institute of Technology Patna via Swayam