Is GPL the Future of Sentence Transformers - Generative Pseudo-Labeling Deep Dive
Offered By: James Briggs via YouTube
Course Description
Overview
Dive deep into Generative Pseudo-Labeling (GPL) and its potential impact on sentence transformers in this comprehensive video tutorial. Explore the challenges of training sentence transformers and how GPL offers a promising solution for fine-tuning high-performance bi-encoder models using unlabeled text data. Learn about the core concepts of GPL, including query generation, negative mining, and pseudo-labeling, with practical code examples using the CORD-19 dataset. Discover the importance of these techniques in building intelligent language models capable of understanding and responding to natural language queries. Gain insights into the implementation of GPL, including the use of Margin MSE Loss and fine-tuning strategies. Conclude with a discussion on the future of sentence transformers and the potential applications of GPL across various industries.
Syllabus
Intro
Semantic Web and Other Uses
Why GPL?
How GPL Works
Query Generation
CORD-19 Dataset and Download
Query Generation Code
Query Generation is Not Perfect
Negative Mining
Negative Mining Implementation
Negative Mining Code
Pseudo-Labeling
Pseudo-Labeling Code
Importance of Pseudo-Labeling
Margin MSE Loss
MarginMSE Fine-tune Code
Choosing Number of Steps
Fast Evaluation
What's Next for Sentence Transformers?
Taught by
James Briggs
Related Courses
Building a unique NLP project: 1984 book vs 1984 albumCoursera Project Network via Coursera Exam Prep AI-102: Microsoft Azure AI Engineer Associate
Whizlabs via Coursera Amazon Echo Reviews Sentiment Analysis Using NLP
Coursera Project Network via Coursera Amazon Translate: Translate documents with batch translation
Coursera Project Network via Coursera Analyze Text Data with Yellowbrick
Coursera Project Network via Coursera