Shaping the Future of AI from the History of Transformer Architectures - Stanford CS25
Offered By: Stanford University via YouTube
Course Description
Overview
Explore the evolution of Transformer architectures and their impact on AI development in this insightful lecture by OpenAI research scientist Hyung Won Chung. Gain a unique perspective on the driving forces behind AI advancements, focusing on the role of exponentially cheaper compute and associated scaling. Examine the early history of Transformer architectures, understanding the motivations behind each development and how they became less relevant with increased computational power. Connect past and present AI trends to project future directions in the field. Delve into the differences between encoder-decoder and decoder-only models, and learn about the rationale for encoder-decoder's additional structures from a scaling perspective. Benefit from Chung's extensive experience in Large Language Models, including work on pre-training, instruction fine-tuning, reinforcement learning with human feedback, reasoning, and multilinguality.
Syllabus
Stanford CS25: V4 I Hyung Won Chung of OpenAI
Taught by
Stanford Online
Tags
Related Courses
Artificial Intelligence Foundations: Neural NetworksLinkedIn Learning Transformers: Text Classification for NLP Using BERT
LinkedIn Learning TensorFlow: Working with NLP
LinkedIn Learning BERTによる自然言語処理を学ぼう! -Attention、TransformerからBERTへとつながるNLP技術-
Udemy Complete Natural Language Processing Tutorial in Python
Keith Galli via YouTube