Understanding the LLM Economics: The $360k Question - Lecture
Offered By: MLOps.community via YouTube
Course Description
Overview
Explore the economics of Large Language Models (LLMs) in production through this insightful conference talk from the LLMs in Production Conference. Dive deep into the costs involved in building LLM-based applications, comparing expenses for RAG versus fine-tuning approaches and open-source versus commercial LLMs. Discover eye-opening examples, such as the $360,000 price tag for summarizing Wikipedia using GPT-4's 8k context window. Gain valuable insights into optimizing LLM costs, understanding the trade-offs between different approaches, and learn strategies for maintaining cost-effectiveness as LLM applications move beyond the honeymoon phase into practical realities of production environments.
Syllabus
Intro
Presentation
Introduction
Goal of the talk
Math Presentation
Problem Statement
Disclaimer
GPT4 Model
Selfhosted models
Fine tuning
OpenAI Fine tuning
Key takeaways
Moveworks example
Open source vs commercial
Offloading tasks
True Foundry
Total Cost
Lossless Compression
Open Source Models
Outro
Taught by
MLOps.community
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Natural Language Processing
Columbia University via Coursera Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent