YoVDO

Shaping the Future of AI from the History of Transformer Architectures - Stanford CS25

Offered By: Stanford University via YouTube

Tags

Transformer Architecture Courses Encoder-Decoder Models Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the evolution of Transformer architectures and their impact on AI development in this insightful lecture by OpenAI research scientist Hyung Won Chung. Gain a unique perspective on the driving forces behind AI advancements, focusing on the role of exponentially cheaper compute and associated scaling. Examine the early history of Transformer architectures, understanding the motivations behind each development and how they became less relevant with increased computational power. Connect past and present AI trends to project future directions in the field. Delve into the differences between encoder-decoder and decoder-only models, and learn about the rationale for encoder-decoder's additional structures from a scaling perspective. Benefit from Chung's extensive experience in Large Language Models, including work on pre-training, instruction fine-tuning, reinforcement learning with human feedback, reasoning, and multilinguality.

Syllabus

Stanford CS25: V4 I Hyung Won Chung of OpenAI


Taught by

Stanford Online

Tags

Related Courses

Artificial Intelligence Foundations: Neural Networks
LinkedIn Learning
Transformers: Text Classification for NLP Using BERT
LinkedIn Learning
TensorFlow: Working with NLP
LinkedIn Learning
BERTによる自然言語処理を学ぼう! -Attention、TransformerからBERTへとつながるNLP技術-
Udemy
Complete Natural Language Processing Tutorial in Python
Keith Galli via YouTube