YoVDO

Pre-Trained Multilingual Sequence to Sequence Models for NMT - Tips, Tricks and Challenges

Offered By: Toronto Machine Learning Series (TMLS) via YouTube

Tags

Sequence to Sequence Models Courses Pre-trained Models Courses Fine-Tuning Courses Natural Language Inference Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the world of Neural Machine Translation (NMT) in this comprehensive 90-minute tutorial presented by Annie En-Shiun Lee, Assistant Professor at the University of Toronto's Computer Science Department. Delve into the rapid evolution of NMT and the power of Pre-trained Multilingual Sequence to Sequence (PMSS) models like mBART and mT5. Learn how these models, pre-trained on extensive general data, can be fine-tuned for impressive results in various natural language tasks. Gain insights into adapting pre-trained models for NMT, discover essential tips and tricks for training and evaluation, and understand the challenges faced when implementing these models. Whether you're approaching NMT from a research or industry perspective, this tutorial offers valuable knowledge to enhance your understanding and application of cutting-edge translation technology.

Syllabus

Pre-Trained Multilingual Sequence to Sequence Models for NMT Tips, Tricks and Challenges


Taught by

Toronto Machine Learning Series (TMLS)

Related Courses

TensorFlow: Working with NLP
LinkedIn Learning
Introduction to Video Editing - Video Editing Tutorials
Great Learning via YouTube
HuggingFace Crash Course - Sentiment Analysis, Model Hub, Fine Tuning
Python Engineer via YouTube
GPT3 and Finetuning the Core Objective Functions - A Deep Dive
David Shapiro ~ AI via YouTube
How to Build a Q&A AI in Python - Open-Domain Question-Answering
James Briggs via YouTube