YoVDO

Coding a ChatGPT-Like Transformer From Scratch in PyTorch

Offered By: StatQuest with Josh Starmer via YouTube

Tags

PyTorch Courses Machine Learning Courses Deep Learning Courses Neural Networks Courses ChatGPT Courses Transformers Courses Attention Mechanisms Courses Model Training Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Walk through the process of coding a ChatGPT-like Transformer from scratch using PyTorch in this comprehensive 31-minute video tutorial. Learn how to load necessary modules, create a training dataset, implement position encoding, code attention mechanisms, and build a decoder-only Transformer. Observe the model running untrained before diving into the training process and practical application. Gain insights into the step-by-step implementation with clear explanations of every detail, assuming prior knowledge of decoder-only Transformers, essential matrix algebra for neural networks, and matrix math behind Transformers.

Syllabus

Awesome song and introduction
Loading the modules
Creating the training dataset
Coding Position Encoding
Coding Attention
Coding a Decoder-Only Transformer
Running the model untrained
Training and using the model


Taught by

StatQuest with Josh Starmer

Related Courses

Neural Networks for Machine Learning
University of Toronto via Coursera
機器學習技法 (Machine Learning Techniques)
National Taiwan University via Coursera
Machine Learning Capstone: An Intelligent Application with Deep Learning
University of Washington via Coursera
Прикладные задачи анализа данных
Moscow Institute of Physics and Technology via Coursera
Leading Ambitious Teaching and Learning
Microsoft via edX