Fine-tuning Methods for Vector Search in Semantic Search and QA Applications
Offered By: OpenSource Connections via YouTube
Course Description
Overview
Explore fine-tuning techniques for vector search in this 36-minute conference talk from Haystack EU 2022. Delve into the challenges of building effective embedding models for domain-specific applications. Learn about popular fine-tuning methods for semantic search and QA, including MSE-loss, MNR-loss, multilingual knowledge distillation, TSDAE, AugSBERT, GenQ, and GPL. Understand when and how to apply these techniques based on available data and use cases. Gain insights from James Briggs, a Staff Developer Advocate at Pinecone and freelance ML Engineer, as he shares his expertise in NLP and vector search. Discover strategies for handling low-resource scenarios, unstructured text, and data augmentation techniques to improve your embedding models.
Syllabus
Intro
Welcome
Vector Search
Why finetune
What is finetuning
Multiple and exit ranking
Hard negative mining
How many pairs
Low resource scenarios
Unstructured text
Synthetic data augmentation
Asymmetric data augmentation
Taught by
OpenSource Connections
Related Courses
U&P AI - Natural Language Processing (NLP) with PythonUdemy What's New in Cognitive Search and Cool Frameworks with PyTorch - Episode 5
Microsoft via YouTube Stress Testing Qdrant - Semantic Search with 90,000 Vectors - Lightning Fast Search Microservice
David Shapiro ~ AI via YouTube Semantic Search for AI - Testing Out Qdrant Neural Search
David Shapiro ~ AI via YouTube Spotify's Podcast Search Explained
James Briggs via YouTube