Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Pre-Trained Multilingual Sequence to Sequence Models for NMT - Tips, Tricks and Challenges

Toronto Machine Learning Series (TMLS) via YouTube

Overview

Explore the world of Neural Machine Translation (NMT) in this comprehensive 90-minute tutorial presented by Annie En-Shiun Lee, Assistant Professor at the University of Toronto's Computer Science Department. Delve into the rapid evolution of NMT and the power of Pre-trained Multilingual Sequence to Sequence (PMSS) models like mBART and mT5. Learn how these models, pre-trained on extensive general data, can be fine-tuned for impressive results in various natural language tasks. Gain insights into adapting pre-trained models for NMT, discover essential tips and tricks for training and evaluation, and understand the challenges faced when implementing these models. Whether you're approaching NMT from a research or industry perspective, this tutorial offers valuable knowledge to enhance your understanding and application of cutting-edge translation technology.

Syllabus

Pre-Trained Multilingual Sequence to Sequence Models for NMT Tips, Tricks and Challenges

Taught by

Toronto Machine Learning Series (TMLS)

Reviews

Start your review of Pre-Trained Multilingual Sequence to Sequence Models for NMT - Tips, Tricks and Challenges

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.