YoVDO

Serverless for ML Inference on Kubernetes - Panacea or Folly

Offered By: CNCF [Cloud Native Computing Foundation] via YouTube

Tags

Conference Talks Courses Kubernetes Courses Cost Analysis Courses System Architecture Courses Serverless Computing Courses Usability Courses Machine Learning Inference Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the advantages and challenges of serverless computing for machine learning inference on Kubernetes in this insightful conference talk. Delve into the results of extensive benchmarking experiments comparing serverless and traditional computing for inference workloads running on Kubernetes, using KubeFlow and the ModelDB MLOps Toolkit. Gain valuable insights into various model types, data modalities, hardware configurations, and workloads. Learn how to architect your own Kubernetes-based ML inference system and understand the trade-offs between flexibility, operating costs, and performance. Discover whether serverless computing is truly a panacea for elastic compute in ML inference or if its limitations outweigh its benefits.

Syllabus

Introduction
What is Serverless
ML Serving Considerations
Benchmark
Usability
Cost
Summary


Taught by

CNCF [Cloud Native Computing Foundation]

Related Courses

Introduction to AWS Inferentia and Amazon EC2 Inf1 Instances
Pluralsight
Introduction to AWS Inferentia and Amazon EC2 Inf1 Instances (Korean)
Amazon Web Services via AWS Skill Builder
Introduction to Amazon Elastic Inference
Amazon Web Services via AWS Skill Builder
TensorFlow Lite - Solution for Running ML On-Device
TensorFlow via YouTube
Inference on KubeEdge
Linux Foundation via YouTube