The Software GPU - Making Inference Scale in the Real World
Offered By: Open Data Science via YouTube
Course Description
Overview
Discover how to process complex deep learning models at scale without specialized AI hardware in this 32-minute conference talk. Explore the Neural Magic Inference Engine, which challenges conventional wisdom about high-throughput computing by utilizing commodity CPUs instead of GPUs. Learn how CPUs' sophisticated cache memory hierarchies can address the limitations of GPUs, allowing for exceptional model performance without sacrificing accuracy. Gain insights into running deep neural networks with GPU-class performance on commodity CPUs, offering deployment flexibility and cost-effectiveness for data science teams. Witness a live demonstration of the Neural Magic Demo and understand how this software-based approach may revolutionize machine learning inference in real-world applications.
Syllabus
Intro
1/100 of a second
Neural Tissue
Compute
Image Recognition
Memory Size
Brain in Silicon
But We Are Learning
Efficient Net-B4 (AmoebaNet Accuracy)
Future of Neural Hardware/Software
CPU vs. GPU
ResNet50 87.5% Sparse (1% Acc Loss)
A Software GPU
Live Webinar: Neural Magic Demo
Taught by
Open Data Science
Related Courses
Neural Networks for Machine LearningUniversity of Toronto via Coursera Good Brain, Bad Brain: Basics
University of Birmingham via FutureLearn Statistical Learning with R
Stanford University via edX Machine Learning 1—Supervised Learning
Brown University via Udacity Fundamentals of Neuroscience, Part 2: Neurons and Networks
Harvard University via edX