YoVDO

Running ML Inference Services in Shared Hosting Environments

Offered By: MLOps World: Machine Learning in Production via YouTube

Tags

Machine Learning Courses Kubernetes Courses MLOps Courses OpenMP Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the challenges and solutions of running machine learning inference services in shared hosting environments like ECS and Kubernetes in this 26-minute conference talk from MLOps World. Learn how Nextdoor's ML team identified and resolved issues affecting latency and throughput, resulting in significant performance improvements. Discover key insights on request queue management and OpenMP parameter tuning to optimize ML inference services. Gain valuable knowledge from machine learning engineer Danny Luo's experience in implementing ML solutions in complex enterprise environments, and understand how to achieve substantial latency reductions, throughput increases, and improved resource utilization while maintaining performance in shared hosting setups.

Syllabus

Running ML Inference Services in Shared Hosting Environments


Taught by

MLOps World: Machine Learning in Production

Related Courses

Machine Learning Operations (MLOps): Getting Started
Google Cloud via Coursera
Проектирование и реализация систем машинного обучения
Higher School of Economics via Coursera
Demystifying Machine Learning Operations (MLOps)
Pluralsight
Machine Learning Engineer with Microsoft Azure
Microsoft via Udacity
Machine Learning Engineering for Production (MLOps)
DeepLearning.AI via Coursera