How to Build a Bert WordPiece Tokenizer in Python and HuggingFace
Offered By: James Briggs via YouTube
Course Description
Overview
Learn how to build a BERT WordPiece tokenizer from scratch using Python and HuggingFace in this comprehensive tutorial video. Explore the process of creating a custom tokenizer for specific use cases, particularly for uncommon languages or less tech-savvy domains. Dive into the intricacies of the WordPiece tokenizer used by BERT, a popular transformer model for various language-based machine learning tasks. Follow along as the instructor guides you through downloading datasets, utilizing HuggingFace's tools, and implementing the tokenizer code. Gain valuable insights into the tokenizer walkthrough and understand how this fundamental step can enhance your natural language processing projects.
Syllabus
Intro
WordPiece Tokenizer
Download Data Sets
HuggingFace
Dataset
Tokenizer
Tokenizer Walkthrough
Tokenizer Code
Taught by
James Briggs
Related Courses
Sequence ModelsDeepLearning.AI via Coursera Modern Natural Language Processing in Python
Udemy Stanford Seminar - Transformers in Language: The Development of GPT Models Including GPT-3
Stanford University via YouTube Long Form Question Answering in Haystack
James Briggs via YouTube Spotify's Podcast Search Explained
James Briggs via YouTube