YoVDO

Monitoring ML Models - Full Stack Deep Learning - Spring 2021

Offered By: The Full Stack via YouTube

Tags

Deep Learning Courses Data Drift Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn how to monitor machine learning models in production and keep them healthy in this 37-minute lecture from the Full Stack Deep Learning Spring 2021 series. Explore the reasons behind model performance degradation post-deployment, understand data drift, and discover what aspects of your models to monitor. Gain insights into measuring changes, determining if changes are detrimental, and familiarize yourself with monitoring tools. Examine the relationship between monitoring and your broader ML system, and conclude with key takeaways for maintaining optimal model performance in real-world applications.

Syllabus

​ - Introduction
​ - Model Performance Degrades Post-Deployment
​ - Data Drift
​ - What To Monitor?
​ - How To Measure When Things Change
​ - How To Tell If A Change Is Bad
​ - Tools For Monitoring
​ - Monitoring And Your Broader ML System
- Takeaways


Taught by

The Full Stack

Related Courses

How to Detect Silent Failures in ML Models
Data Science Dojo via YouTube
Dataset Management for Computer Vision - Important Component to Delivering Computer Vision Solutions
Open Data Science via YouTube
Testing ML Models in Production - Detecting Data and Concept Drift
Databricks via YouTube
Ekya - Continuous Learning of Video Analytics Models on Edge Compute Servers
USENIX via YouTube
Building and Maintaining High-Performance AI
Data Science Dojo via YouTube