Is GPL the Future of Sentence Transformers - Generative Pseudo-Labeling Deep Dive
Offered By: James Briggs via YouTube
Course Description
Overview
Dive deep into Generative Pseudo-Labeling (GPL) and its potential impact on sentence transformers in this comprehensive video tutorial. Explore the challenges of training sentence transformers and how GPL offers a promising solution for fine-tuning high-performance bi-encoder models using unlabeled text data. Learn about the core concepts of GPL, including query generation, negative mining, and pseudo-labeling, with practical code examples using the CORD-19 dataset. Discover the importance of these techniques in building intelligent language models capable of understanding and responding to natural language queries. Gain insights into the implementation of GPL, including the use of Margin MSE Loss and fine-tuning strategies. Conclude with a discussion on the future of sentence transformers and the potential applications of GPL across various industries.
Syllabus
Intro
Semantic Web and Other Uses
Why GPL?
How GPL Works
Query Generation
CORD-19 Dataset and Download
Query Generation Code
Query Generation is Not Perfect
Negative Mining
Negative Mining Implementation
Negative Mining Code
Pseudo-Labeling
Pseudo-Labeling Code
Importance of Pseudo-Labeling
Margin MSE Loss
MarginMSE Fine-tune Code
Choosing Number of Steps
Fast Evaluation
What's Next for Sentence Transformers?
Taught by
James Briggs
Related Courses
Microsoft Bot Framework and Conversation as a PlatformMicrosoft via edX Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube Improving Customer Experiences with Speech to Text and Text to Speech
Microsoft via YouTube Stanford Seminar - Deep Learning in Speech Recognition
Stanford University via YouTube Select Topics in Python: Natural Language Processing
Codio via Coursera