Alpa: Simple Large Model Training and Inference on Ray
Offered By: Anyscale via YouTube
Course Description
Overview
Explore the capabilities of Alpa, a Ray-native library designed for automated training and serving of large models like GPT-3. Discover how Alpa simplifies model-parallel training of complex deep learning models by generating execution plans that unify data, operator, and pipeline parallelism. Learn about Alpa's innovative approach to distributing training across two hierarchical levels of parallelism: inter-operator and intra-operator. Understand how Alpa constructs a new hierarchical space for massive model-parallel execution plans and uses compilation passes to derive optimal parallel execution plans. Examine Alpa's efficient runtime that orchestrates two-level parallel execution on distributed compute devices. Compare Alpa's performance to hand-tuned model-parallel training systems and explore its versatility in handling models with heterogeneous architectures. Delve into both the algorithmic aspects and the engineering/system implementation, with a focus on Ray's crucial role as a building block of the Alpa runtime. This 31-minute talk from Anyscale at Ray Summit provides valuable insights into advanced techniques for scaling out complex deep learning models on distributed computing environments.
Syllabus
Alpa - Simple large model training and inference on Ray
Taught by
Anyscale
Related Courses
How to Build Codex SolutionsMicrosoft via YouTube Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube Building Intelligent Applications with World-Class AI
Microsoft via YouTube Stanford Seminar - Transformers in Language: The Development of GPT Models Including GPT-3
Stanford University via YouTube ChatGPT: GPT-3, GPT-4 Turbo: Unleash the Power of LLM's
Udemy