StreamBox - A Lightweight GPU Sandbox for Serverless Inference Workflow
Offered By: USENIX via YouTube
Course Description
Overview
Explore a groundbreaking conference talk on StreamBox, a lightweight GPU sandbox designed for serverless inference workflows. Delve into the challenges of dynamic workloads and latency-sensitive DNN inference in serverless computing environments. Discover how StreamBox addresses the limitations of existing serverless inference systems by implementing fine-grained and auto-scaling memory management, enabling transparent and efficient intra-GPU communication across functions, and facilitating PCIe bandwidth sharing among concurrent streams. Learn about the significant improvements StreamBox offers, including up to 82% reduction in GPU memory footprint and a 6.7X increase in throughput compared to state-of-the-art systems. Gain insights into the potential impact of this innovative approach on scalable DNN inference serving and the future of serverless computing for GPU-intensive tasks.
Syllabus
USENIX ATC '24 - StreamBox: A Lightweight GPU SandBox for Serverless Inference Workflow
Taught by
USENIX
Related Courses
High Performance ComputingGeorgia Institute of Technology via Udacity Fundamentals of Accelerated Computing with CUDA C/C++
Nvidia via Independent High Performance Computing for Scientists and Engineers
Indian Institute of Technology, Kharagpur via Swayam CUDA programming Masterclass with C++
Udemy Neural Network Programming - Deep Learning with PyTorch
YouTube