Boosting LLM Development and Training Efficiency: Automated Parallelization with MindSpore
Offered By: Linux Foundation via YouTube
Course Description
Overview
Explore automated parallelization techniques for boosting large language model (LLM) development and training efficiency using MindSpore in this 44-minute conference talk by Yiren Xing from Huawei. Learn about an innovative approach that allows developers to focus on algorithm research without the need for intrusive model modifications. Discover how distributed training on large-scale clusters can be achieved through simple strategy configurations. Gain insights into MindSpore's hyperparameter search model for automatically finding optimal parallelization strategies, which can achieve 90%-110% of expert tuning performance. Understand how this method significantly reduces model modification time while efficiently accelerating LLM training. The presentation covers both the challenges of large-scale distributed parallel training and the solutions offered by automated parallelization, making it valuable for AI researchers and developers working with large language models.
Syllabus
Boosting LLM Development & Training Efficiency: Automated Parallelization with MindSpore- Yiren Xing
Taught by
Linux Foundation
Tags
Related Courses
Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and OptimizationDeepLearning.AI via Coursera How to Win a Data Science Competition: Learn from Top Kagglers
Higher School of Economics via Coursera Predictive Modeling and Machine Learning with MATLAB
MathWorks via Coursera Machine Learning Rapid Prototyping with IBM Watson Studio
IBM via Coursera Hyperparameter Tuning with Neural Network Intelligence
Coursera Project Network via Coursera