Making the Most Out of Hardware Accelerators in Kubernetes Clusters
Offered By: CNCF [Cloud Native Computing Foundation] via YouTube
Course Description
Overview
Explore techniques for optimizing hardware accelerators in Kubernetes clusters for MLOps in this 25-minute conference talk. Learn how to efficiently manage GPUs and TPUs using open-source projects like Kubeflow, TensorFlow, and Kubernetes. Discover strategies for differentiating between compute instances and hardware accelerators, determining appropriate Pod placement, scaling, time-sharing accelerators, and establishing effective compute-accelerator links. Gain valuable insights to enhance machine learning workload performance and maximize the potential of your hardware resources in a Kubernetes environment.
Syllabus
Making the Most Out of Your Hardware Accelerators in a Kubernetes Clu... Rishit Dagli & Shivay Lamba
Taught by
CNCF [Cloud Native Computing Foundation]
Related Courses
Building End-to-end Machine Learning Workflows with KubeflowPluralsight Smart Analytics, Machine Learning, and AI on GCP
Pluralsight Leveraging Cloud-Based Machine Learning on Google Cloud Platform: Real World Applications
LinkedIn Learning Distributed TensorFlow - TensorFlow at O'Reilly AI Conference, San Francisco '18
TensorFlow via YouTube KFServing - Model Monitoring with Apache Spark and Feature Store
Databricks via YouTube