Distributed Training: Techniques and Strategies - Part II - Lecture 18
Offered By: MIT HAN Lab via YouTube
Course Description
Overview
Dive into advanced concepts of distributed training in machine learning with this 55-minute lecture from MIT's 6.5940 course. Explore key topics in the second part of distributed training, presented by Prof. Song Han from the MIT HAN Lab. Gain insights into cutting-edge techniques for scaling machine learning models across multiple devices and systems. Access accompanying slides at efficientml.ai to enhance your understanding of the material covered in this comprehensive session on efficient machine learning practices.
Syllabus
EfficientML.ai Lecture 18: Distributed Training (Part II) (MIT 6.5940, Fall 2023)
Taught by
MIT HAN Lab
Related Courses
Custom and Distributed Training with TensorFlowDeepLearning.AI via Coursera Architecting Production-ready ML Models Using Google Cloud ML Engine
Pluralsight Building End-to-end Machine Learning Workflows with Kubeflow
Pluralsight Deploying PyTorch Models in Production: PyTorch Playbook
Pluralsight Inside TensorFlow
TensorFlow via YouTube