YoVDO

Alpa: Simple Large Model Training and Inference on Ray

Offered By: Anyscale via YouTube

Tags

Deep Learning Courses GPT-3 Courses Distributed Computing Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the capabilities of Alpa, a Ray-native library designed for automated training and serving of large models like GPT-3. Discover how Alpa simplifies model-parallel training of complex deep learning models by generating execution plans that unify data, operator, and pipeline parallelism. Learn about Alpa's innovative approach to distributing training across two hierarchical levels of parallelism: inter-operator and intra-operator. Understand how Alpa constructs a new hierarchical space for massive model-parallel execution plans and uses compilation passes to derive optimal parallel execution plans. Examine Alpa's efficient runtime that orchestrates two-level parallel execution on distributed compute devices. Compare Alpa's performance to hand-tuned model-parallel training systems and explore its versatility in handling models with heterogeneous architectures. Delve into both the algorithmic aspects and the engineering/system implementation, with a focus on Ray's crucial role as a building block of the Alpa runtime. This 31-minute talk from Anyscale at Ray Summit provides valuable insights into advanced techniques for scaling out complex deep learning models on distributed computing environments.

Syllabus

Alpa - Simple large model training and inference on Ray


Taught by

Anyscale

Related Courses

How to Build Codex Solutions
Microsoft via YouTube
Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube
Building Intelligent Applications with World-Class AI
Microsoft via YouTube
Stanford Seminar - Transformers in Language: The Development of GPT Models Including GPT-3
Stanford University via YouTube
ChatGPT: GPT-3, GPT-4 Turbo: Unleash the Power of LLM's
Udemy