YoVDO

Bernice: A Multilingual Pre-trained Encoder for Twitter

Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube

Tags

Data Mining Courses Machine Learning Courses Twitter Courses Computational Linguistics Courses Text Classification Courses Language Models Courses RoBERTa Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a groundbreaking multilingual RoBERTa language model called Bernice, designed specifically for Twitter data analysis. Learn about the development of this pre-trained encoder, which was trained from scratch on 2.5 billion tweets across multiple languages. Discover how Bernice outperforms other models adapted to social media data and strong multilingual baselines in various monolingual and multilingual Twitter benchmarks. Gain insights into the unique challenges of processing Twitter's multilingual content and how Bernice addresses the significant differences between Twitter language and other domains commonly used in large language model training.

Syllabus

Bernice: A Multilingual Pre-trained Encoder for Twitter - Alexandra DeLucia - October 2022


Taught by

Center for Language & Speech Processing(CLSP), JHU

Related Courses

Microsoft Bot Framework and Conversation as a Platform
Microsoft via edX
Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube
Improving Customer Experiences with Speech to Text and Text to Speech
Microsoft via YouTube
Stanford Seminar - Deep Learning in Speech Recognition
Stanford University via YouTube
Select Topics in Python: Natural Language Processing
Codio via Coursera