YoVDO

Fine-tuning Methods for Vector Search in Semantic Search and QA Applications

Offered By: OpenSource Connections via YouTube

Tags

Vector Search Courses Semantic Search Courses Fine-Tuning Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore fine-tuning techniques for vector search in this 36-minute conference talk from Haystack EU 2022. Delve into the challenges of building effective embedding models for domain-specific applications. Learn about popular fine-tuning methods for semantic search and QA, including MSE-loss, MNR-loss, multilingual knowledge distillation, TSDAE, AugSBERT, GenQ, and GPL. Understand when and how to apply these techniques based on available data and use cases. Gain insights from James Briggs, a Staff Developer Advocate at Pinecone and freelance ML Engineer, as he shares his expertise in NLP and vector search. Discover strategies for handling low-resource scenarios, unstructured text, and data augmentation techniques to improve your embedding models.

Syllabus

Intro
Welcome
Vector Search
Why finetune
What is finetuning
Multiple and exit ranking
Hard negative mining
How many pairs
Low resource scenarios
Unstructured text
Synthetic data augmentation
Asymmetric data augmentation


Taught by

OpenSource Connections

Related Courses

U&P AI - Natural Language Processing (NLP) with Python
Udemy
What's New in Cognitive Search and Cool Frameworks with PyTorch - Episode 5
Microsoft via YouTube
Stress Testing Qdrant - Semantic Search with 90,000 Vectors - Lightning Fast Search Microservice
David Shapiro ~ AI via YouTube
Semantic Search for AI - Testing Out Qdrant Neural Search
David Shapiro ~ AI via YouTube
Spotify's Podcast Search Explained
James Briggs via YouTube