YoVDO

Pre-Trained Multilingual Sequence to Sequence Models for NMT - Tips, Tricks and Challenges

Offered By: Toronto Machine Learning Series (TMLS) via YouTube

Tags

Sequence to Sequence Models Courses Pre-trained Models Courses Fine-Tuning Courses Natural Language Inference Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the world of Neural Machine Translation (NMT) in this comprehensive 90-minute tutorial presented by Annie En-Shiun Lee, Assistant Professor at the University of Toronto's Computer Science Department. Delve into the rapid evolution of NMT and the power of Pre-trained Multilingual Sequence to Sequence (PMSS) models like mBART and mT5. Learn how these models, pre-trained on extensive general data, can be fine-tuned for impressive results in various natural language tasks. Gain insights into adapting pre-trained models for NMT, discover essential tips and tricks for training and evaluation, and understand the challenges faced when implementing these models. Whether you're approaching NMT from a research or industry perspective, this tutorial offers valuable knowledge to enhance your understanding and application of cutting-edge translation technology.

Syllabus

Pre-Trained Multilingual Sequence to Sequence Models for NMT Tips, Tricks and Challenges


Taught by

Toronto Machine Learning Series (TMLS)

Related Courses

Perform Real-Time Object Detection with YOLOv3
Coursera Project Network via Coursera
Intel® Edge AI Fundamentals with OpenVINO™
Intel via Udacity
Building Deep Learning Applications with Keras 2.0
LinkedIn Learning
Expediting Deep Learning with Transfer Learning: PyTorch Playbook
Pluralsight
2024 Introduction to Spacy for Natural Language Processing
Udemy