YoVDO

Shaping the Future of AI from the History of Transformer Architectures - Stanford CS25

Offered By: Stanford University via YouTube

Tags

Transformer Architecture Courses Encoder-Decoder Models Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the evolution of Transformer architectures and their impact on AI development in this insightful lecture by OpenAI research scientist Hyung Won Chung. Gain a unique perspective on the driving forces behind AI advancements, focusing on the role of exponentially cheaper compute and associated scaling. Examine the early history of Transformer architectures, understanding the motivations behind each development and how they became less relevant with increased computational power. Connect past and present AI trends to project future directions in the field. Delve into the differences between encoder-decoder and decoder-only models, and learn about the rationale for encoder-decoder's additional structures from a scaling perspective. Benefit from Chung's extensive experience in Large Language Models, including work on pre-training, instruction fine-tuning, reinforcement learning with human feedback, reasoning, and multilinguality.

Syllabus

Stanford CS25: V4 I Hyung Won Chung of OpenAI


Taught by

Stanford Online

Tags

Related Courses

Natural Language Processing on Google Cloud
Google Cloud via Coursera
MIT 6.S191 - Automatic Speech Recognition
Alexander Amini via YouTube
Introduction to T5 for Sentiment Span Extraction
Abhishek Thakur via YouTube
CMU Advanced NLP 2021 - Conditional Generation
Graham Neubig via YouTube
Neural Nets for NLP 2021 - Conditioned Generation
Graham Neubig via YouTube