YoVDO

LLaMA- Open and Efficient Foundation Language Models - Paper Explained

Offered By: Yannic Kilcher via YouTube

Tags

LLaMA (Large Language Model Meta AI) Courses Hyperparameters Courses

Course Description

Overview

Explore an in-depth analysis of Meta AI's LLaMA, a series of large language models ranging from 7B to 65B parameters. Dive into the technical aspects of this groundbreaking research, including training data, hyperparameters, architecture modifications, and efficient implementation. Learn how LLaMA outperforms GPT-3 on most benchmarks despite being significantly smaller. Understand the implications of open-sourcing such models for the research community and the potential impact on the field of artificial intelligence. Gain insights into the main results, model completions, and the future of foundation language models.

Syllabus

- Introduction & Paper Overview
- Rant on Open-Sourcing
- Training Data
- Training Hyperparameters
- Architecture Modifications
- Optimizer
- Efficient Implementation
- Main Results
- Some more completions
- Conclusion


Taught by

Yannic Kilcher

Related Courses

Art and Science of Machine Learning em Português Brasileiro
Google Cloud via Coursera
Data Science: Supervised Machine Learning in Python
Udemy
Machine Learning - Regression and Classification (math Inc.)
Udemy
Artificial Neural Networks(ANN) Made Easy
Udemy
Art and Science of Machine Learning
Pluralsight