Using Vector Databases: Practical Advice for Production - LLMs in Prod Conference
Offered By: MLOps.community via YouTube
Course Description
Overview
Discover practical advice for using vector databases in production environments during this 30-minute conference talk from the LLMs in Prod Conference. Explore various use cases for vector databases with large language models, including information retrieval, conversational memory for chatbots, and semantic caching. Delve into the less flashy but crucial aspects of implementing these technologies, such as prompt engineering, text chunking, compliance considerations, on-premise solutions, embedding model changes, index types, A/B testing, cloud platform selection, deployment strategies, feature injection, and tool comparisons. Gain insights from a year's worth of Redis deployments for AI use cases, condensed into a comprehensive overview. Learn about similarity searches, design patterns, context retrieval, feature injection, query optimization, guard rails, long-term memory, common mistakes, and index management. Benefit from the expertise of Sam Partee, a Principal Engineer guiding AI efforts at Redis, as he shares valuable knowledge on integrating vector databases into ML pipelines for feature storage, search, and inference workloads.
Syllabus
Intro
Similarity searches
Redis
Design Patterns
Context Retrieval
Hide
Feature Injection
Query Get
Guard Rails
Longterm Memory
Common Mistakes
Laziness
Index Management
Metadata
Taught by
MLOps.community
Related Courses
Vector Similarity SearchData Science Dojo via YouTube Supercharging Semantic Search with Pinecone and Cohere
Pinecone via YouTube Search Like You Mean It - Semantic Search with NLP and a Vector Database
Pinecone via YouTube The Rise of Vector Data
Pinecone via YouTube NER Powered Semantic Search in Python
James Briggs via YouTube