LLaMA- Open and Efficient Foundation Language Models - Paper Explained
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore an in-depth analysis of Meta AI's LLaMA, a series of large language models ranging from 7B to 65B parameters. Dive into the technical aspects of this groundbreaking research, including training data, hyperparameters, architecture modifications, and efficient implementation. Learn how LLaMA outperforms GPT-3 on most benchmarks despite being significantly smaller. Understand the implications of open-sourcing such models for the research community and the potential impact on the field of artificial intelligence. Gain insights into the main results, model completions, and the future of foundation language models.
Syllabus
- Introduction & Paper Overview
- Rant on Open-Sourcing
- Training Data
- Training Hyperparameters
- Architecture Modifications
- Optimizer
- Efficient Implementation
- Main Results
- Some more completions
- Conclusion
Taught by
Yannic Kilcher
Related Courses
Art and Science of Machine Learning em Português BrasileiroGoogle Cloud via Coursera Data Science: Supervised Machine Learning in Python
Udemy Machine Learning - Regression and Classification (math Inc.)
Udemy Artificial Neural Networks(ANN) Made Easy
Udemy Art and Science of Machine Learning
Pluralsight