Enabling Elastic Inference on Edge With Knative and EDL
Offered By: CNCF [Cloud Native Computing Foundation] via YouTube
Course Description
Overview
Explore a conference talk on enabling elastic inference on edge using Knative and EDL, presented by Ti Zhou and Daxiang Dong from Baidu. Learn how Baidu leverages these technologies to deploy recommendation model inference on distributed CDN nodes, resulting in a 25% reduction in core network load and a 40% decrease in network communication latency. Discover the experiences and lessons learned from optimizing edge deployment, including auto-scaling, ingress traffic optimization based on geographic location, cold start optimization with model pre-loading, and legacy services interaction with Cloud Events. Gain insights into PaddlePaddle, large-scale training, Elastic Deep Learning (EDL), fault-tolerant training, and elastic knowledge distillation. Understand the benefits of Knative, including its serving capabilities for blue-green deployments, ingress, autoscaling, events, and observability features.
Syllabus
Intro
What is PaddlePaddle?
Large Scale Training with Paddle
Industrial Demand for Elasticity
Elastic Deep Learning (EDL)
Fault Tolerant Training
Elastic Knowledge Distillation
Al Stack
Why Knative
Knative Serving - Blue Green
Knative Serving - Ingress
Knative Serving - Autoscaler
Knative Events
Knative Observability
Taught by
CNCF [Cloud Native Computing Foundation]
Related Courses
Kubernetes EssentialsYouTube Cloud Native Explained
IBM via YouTube Getting Started with Knative
Pluralsight Serverless on Kubernetes Using Knative - Mirantis Labs Tech Talks
Mirantis via YouTube Serverless Containers with Knative and Cloud Run
NDC Conferences via YouTube