LLaMA 3 Deep Dive - Synthetic Data, Privacy, and Model Architecture
Offered By: Aleksa Gordić - The AI Epiphany via YouTube
Course Description
Overview
Dive deep into the latest developments of LLaMA 3 in this informative video featuring Thomas Scialom from Meta. Explore key topics including synthetic data for pre/post training, privacy concerns, scaling and distillation techniques, and the decision not to use Mixture of Experts (MoE) architecture. Learn about the potential upper boundaries for synthetic data quality, context length improvements, Meta's framework choices, and the multilingual capabilities of smaller LLaMA models. Gain valuable insights into the cutting-edge advancements in large language models and their implications for AI research and development.
Syllabus
00:00 - Intro
00:27 - Hyperstack GPUs platform! sponsored
02:08 - What is new in new Llama?
06:40 - Synthetic data
13:30 - Privacy - training on Facebook user data?
15:35 - Scaling and distillation
19:10 - MoE, new architectures?
25:35 - Upper boundary for the quality of SX data?
37:15 - Context length
45:10 - What framework does Meta use for Llama
46:40 - Playing with smaller Llamas
51:20 - Multilingual capabilities
Taught by
Aleksa Gordić - The AI Epiphany
Related Courses
BLEURT - Learning Robust Metrics for Text GenerationYannic Kilcher via YouTube 3D Deep Learning for Gaming with Srinath Sridhar and Stanford Artificial Intelligence
Resemble AI via YouTube Deep Learning in Gaming with Idan Beck
Resemble AI via YouTube Preserving Patient Safety as AI Transforms Clinical Care - Curt Langlotz, Stanford University
Alan Turing Institute via YouTube Synthesizing Plausible Privacy-Preserving Location Traces
IEEE via YouTube