YoVDO

Is GPL the Future of Sentence Transformers - Generative Pseudo-Labeling Deep Dive

Offered By: James Briggs via YouTube

Tags

Natural Language Processing (NLP) Courses Semantic Web Courses Language Models Courses Sentence Transformers Courses

Course Description

Overview

Dive deep into Generative Pseudo-Labeling (GPL) and its potential impact on sentence transformers in this comprehensive video tutorial. Explore the challenges of training sentence transformers and how GPL offers a promising solution for fine-tuning high-performance bi-encoder models using unlabeled text data. Learn about the core concepts of GPL, including query generation, negative mining, and pseudo-labeling, with practical code examples using the CORD-19 dataset. Discover the importance of these techniques in building intelligent language models capable of understanding and responding to natural language queries. Gain insights into the implementation of GPL, including the use of Margin MSE Loss and fine-tuning strategies. Conclude with a discussion on the future of sentence transformers and the potential applications of GPL across various industries.

Syllabus

Intro
Semantic Web and Other Uses
Why GPL?
How GPL Works
Query Generation
CORD-19 Dataset and Download
Query Generation Code
Query Generation is Not Perfect
Negative Mining
Negative Mining Implementation
Negative Mining Code
Pseudo-Labeling
Pseudo-Labeling Code
Importance of Pseudo-Labeling
Margin MSE Loss
MarginMSE Fine-tune Code
Choosing Number of Steps
Fast Evaluation
What's Next for Sentence Transformers?


Taught by

James Briggs

Related Courses

Semantic Search for AI - Testing Out Qdrant Neural Search
David Shapiro ~ AI via YouTube
How to Use OpenAI Whisper to Fix YouTube Search
James Briggs via YouTube
Spotify's Podcast Search Explained
James Briggs via YouTube
Train Sentence Transformers by Generating Queries - GenQ
James Briggs via YouTube
Fine-Tune High Performance Sentence Transformers With Multiple Negatives Ranking
James Briggs via YouTube