Enabling Elastic Inference on Edge With Knative and EDL
Offered By: CNCF [Cloud Native Computing Foundation] via YouTube
Course Description
Overview
Explore a conference talk on enabling elastic inference on edge using Knative and EDL, presented by Ti Zhou and Daxiang Dong from Baidu. Learn how Baidu leverages these technologies to deploy recommendation model inference on distributed CDN nodes, resulting in a 25% reduction in core network load and a 40% decrease in network communication latency. Discover the experiences and lessons learned from optimizing edge deployment, including auto-scaling, ingress traffic optimization based on geographic location, cold start optimization with model pre-loading, and legacy services interaction with Cloud Events. Gain insights into PaddlePaddle, large-scale training, Elastic Deep Learning (EDL), fault-tolerant training, and elastic knowledge distillation. Understand the benefits of Knative, including its serving capabilities for blue-green deployments, ingress, autoscaling, events, and observability features.
Syllabus
Intro
What is PaddlePaddle?
Large Scale Training with Paddle
Industrial Demand for Elasticity
Elastic Deep Learning (EDL)
Fault Tolerant Training
Elastic Knowledge Distillation
Al Stack
Why Knative
Knative Serving - Blue Green
Knative Serving - Ingress
Knative Serving - Autoscaler
Knative Events
Knative Observability
Taught by
CNCF [Cloud Native Computing Foundation]
Related Courses
Data Engineering on Google Cloud Platform 日本語版Google Cloud via Coursera Cloud Computing Fundamentals on Alibaba Cloud
Alibaba Cloud Academy via Coursera Launch an auto-scaling AWS EC2 virtual machine
Coursera Project Network via Coursera Cloud Computing With Amazon Web Services
Udemy AWS Certified Solution Architect - Associate 2020
Udemy