Multi-Class Language Classification With BERT in TensorFlow
Offered By: James Briggs via YouTube
Course Description
Overview
Learn how to build a multi-class language classification model using BERT and TensorFlow in this comprehensive 43-minute tutorial. Explore the power of transformers in natural language processing as you work through each step of the process, from data preprocessing to model training and prediction. Follow along with clearly defined chapters for each section, including data input pipeline creation, model definition, and saving/loading techniques. Gain insights into the significance of transformers in deep learning and their dominance in NLP benchmarks. Utilize the HuggingFace transformers library to create an efficient and high-performing solution for multi-class text classification tasks.
Syllabus
Intro
Pulling Data
Preprocessing
Data Input Pipeline
Defining Model
Model Training
Saving and Loading Models
Making Predictions
Taught by
James Briggs
Related Courses
Neural Networks for Machine LearningUniversity of Toronto via Coursera 機器學習技法 (Machine Learning Techniques)
National Taiwan University via Coursera Machine Learning Capstone: An Intelligent Application with Deep Learning
University of Washington via Coursera Прикладные задачи анализа данных
Moscow Institute of Physics and Technology via Coursera Leading Ambitious Teaching and Learning
Microsoft via edX