Stanford Seminar - Dataflow for Convergence of AI and HPC - GroqChip
Offered By: Stanford University via YouTube
Course Description
Overview
Explore the convergence of AI and High-Performance Computing (HPC) through the lens of dataflow architecture in this Stanford seminar. Delve into the novel Groq architecture and tensor streaming processor (TSP), combining traditional dataflow elements with a powerful stream programming model. Discover how over 400,000 arithmetic units and a SIMD spatial microarchitecture efficiently exploit dataflow locality in deep learning models. Learn about the stream programming model, where on-chip functional units consume and produce tensor inputs, chaining outputs to minimize memory access. Understand how deterministic execution and dataflow locality simplify compiler abstraction, enabling efficient orchestration of data and instructions. Examine the extension of this model to distributed scale-out systems, creating a synchronous parallel computer illusion. Explore how the Groq parallelizing compiler leverages this programming model to auto-scale TSPs, facilitating robust numerical computations in production environments.
Syllabus
Introduction
Dennis Axe
Hardware Software Interface
Pipeline
Core Architecture
Superlane Architecture
DomainSpecific Architecture
Data Types
Communication and Computation
Energy Difference
Functional Control Units
Superlane
Vector Processor
Memory System
Switch Execution Module
System Architecture
Topology
Packaging
Network
Normal RDMA
Communication model
Synchronous communication
Taught by
Stanford Online
Tags
Related Courses
Intro to Parallel ProgrammingNvidia via Udacity Introduction to Linear Models and Matrix Algebra
Harvard University via edX Введение в параллельное программирование с использованием OpenMP и MPI
Tomsk State University via Coursera Supercomputing
Partnership for Advanced Computing in Europe via FutureLearn Fundamentals of Parallelism on Intel Architecture
Intel via Coursera