YoVDO

From Idea to Production: AI Infrastructure for Scaling LLM Applications

Offered By: MLOps World: Machine Learning in Production via YouTube

Tags

MLOps Courses Generative AI Courses GPU Computing Courses Scalability Courses Data Pipelines Courses Cost Optimization Courses Retrieval Augmented Generation (RAG) Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore strategies for scaling Large Language Model (LLM) applications from beta to production in this 38-minute conference talk from MLOps World: Machine Learning in Production. Dive into the challenges of building Generative AI and LLM apps, including adapting to new models, managing complex workflows, and integrating prompt management, data pipelines, RAG, cost optimization, and GPU availability. Learn how to design and deploy adaptable LLM applications, build infrastructure ready for the next best model, and navigate the complexities of bringing AI applications to production. Gain insights on smoothly transitioning LLM apps from development to large-scale deployment, addressing the needs of multiple stakeholders and optimizing for performance and cost.

Syllabus

From Idea to Production: AI Infra for Scaling LLM Apps


Taught by

MLOps World: Machine Learning in Production

Related Courses

Better Llama with Retrieval Augmented Generation - RAG
James Briggs via YouTube
Live Code Review - Pinecone Vercel Starter Template and Retrieval Augmented Generation
Pinecone via YouTube
Nvidia's NeMo Guardrails - Full Walkthrough for Chatbots - AI
James Briggs via YouTube
Hugging Face LLMs with SageMaker - RAG with Pinecone
James Briggs via YouTube
Supercharge Your LLM Applications with RAG
Data Science Dojo via YouTube