YoVDO

VeScale - A PyTorch Native LLM Training Framework for Automatic Parallelism

Offered By: CNCF [Cloud Native Computing Foundation] via YouTube

Tags

PyTorch Courses Deep Learning Courses Neural Networks Courses Parallel Computing Courses Distributed Training Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a groundbreaking PyTorch native framework for large language model (LLM) training in this 24-minute conference talk by Hongyu Zhu from ByteDance. Learn about VeScale, a novel solution that combines PyTorch nativeness with automatic parallelism to address the challenges of distributed training for giant LLMs. Discover how this framework prioritizes ease of use, allowing developers to write single-device PyTorch code while automatically parallelizing it into nD parallelism. Gain insights into the importance of PyTorch ecosystem dominance and the necessity of complex nD parallelism for training massive models. Understand the limitations of existing industry-level frameworks and how VeScale aims to overcome them by offering a user-friendly approach to scaling LLM training.

Syllabus

VeScale: A PyTorch Native LLM Training Framework | veScale:一个PyTorch原生LLM训练框架 - Hongyu Zhu


Taught by

CNCF [Cloud Native Computing Foundation]

Related Courses

Neural Networks for Machine Learning
University of Toronto via Coursera
機器學習技法 (Machine Learning Techniques)
National Taiwan University via Coursera
Machine Learning Capstone: An Intelligent Application with Deep Learning
University of Washington via Coursera
Прикладные задачи анализа данных
Moscow Institute of Physics and Technology via Coursera
Leading Ambitious Teaching and Learning
Microsoft via edX