BGL - GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing
Offered By: USENIX via YouTube
Course Description
Overview
Explore a conference talk that delves into BGL, a distributed GNN training system designed to optimize GPU efficiency for large-scale graph data processing. Learn about innovative techniques for minimizing feature retrieval traffic, improving graph partition algorithms, and reducing cross-partition communication during subgraph sampling. Discover how BGL addresses bottlenecks in preparing data for GPUs, focusing on subgraph sampling and feature retrieval. Gain insights into the dynamic cache engine, co-designed caching policy, and sampling order that achieve a balance between low overhead and high cache hit ratio. Understand the importance of resource isolation in reducing contention between data preprocessing stages. Examine the system's performance improvements over existing GNN training systems, demonstrated through extensive experiments on various GNN models and large graph datasets.
Syllabus
NSDI '23 - BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing
Taught by
USENIX
Related Courses
Scaling Memcache at FacebookUSENIX via YouTube Multi-Person Localization via RF Body Reflections
USENIX via YouTube Opaque - An Oblivious and Encrypted Distributed Analytics Platform
USENIX via YouTube Live Video Analytics at Scale with Approximation and Delay-Tolerance
USENIX via YouTube Clipper - A Low-Latency Online Prediction Serving System
USENIX via YouTube