Enabling Efficient Trillion Parameter Scale Training for Deep Learning Models
Offered By: MLOps.community via YouTube
Course Description
Overview
Explore the challenges and solutions for efficient trillion-parameter scale training in deep learning models in this conference talk from the AI in Production Conference. Delve into DeepSpeed, a deep learning optimization library designed to make distributed model training and inference efficient, effective, and easy on commodity hardware. Learn about training optimizations that improve memory, compute, and data efficiency for extreme model scaling. Gain insights from Olatunji (Tunji) Ruwase, co-founder and lead of the DeepSpeed project at Microsoft, as he shares his expertise in building systems convergence optimizations and frameworks for distributed training and inference of deep learning models.
Syllabus
Enabling Efficient Trillion Parameter Scale Training for Deep Learning Models // Tunji Ruwase
Taught by
MLOps.community
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Artificial Intelligence for Robotics
Stanford University via Udacity Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent