Improving Complex RAG Systems and Achieving No-Regret, Lightning-Fast Deployment Iterations of LLMs
Offered By: Data Science Festival via YouTube
Course Description
Overview
Explore advanced techniques for enhancing complex Retrieval Augmented Generation (RAG) systems and implementing rapid, risk-free deployment iterations of Large Language Models (LLMs) in this 33-minute talk from the Data Science Festival. Delve into strategies to overcome LLM limitations such as outdated training data, restricted context windows, latency, and API rate limits. Learn how to leverage AWS lambda aliases for deploying "shadow" versions and feature flagging beta releases, enabling safe and swift iterations of LLM-based applications. Discover methods to evaluate system performance in production using unique data gathered from these deployment techniques. Gain insights applicable to ML/software engineers and data scientists, with a focus on practical coding experience in LLMs, RAG, and AWS lambdas or equivalent technologies.
Syllabus
Improving complex RAG systems and achieving no regret lightning fast deployment iterations of LLMs
Taught by
Data Science Festival
Related Courses
Pinecone Vercel Starter Template and RAG - Live Code Review Part 2Pinecone via YouTube Will LLMs Kill Search? The Future of Information Retrieval
Aleksa Gordić - The AI Epiphany via YouTube RAG But Better: Rerankers with Cohere AI - Improving Retrieval Pipelines
James Briggs via YouTube Advanced RAG - Contextual Compressors and Filters - Lecture 4
Sam Witteveen via YouTube LangChain Multi-Query Retriever for RAG - Advanced Technique for Broader Vector Space Search
James Briggs via YouTube