YoVDO

CMU Neural Nets for NLP 2018 - Conditioned Generation

Offered By: Graham Neubig via YouTube

Tags

Neural Networks Courses Natural Language Processing (NLP) Courses Language Models Courses

Course Description

Overview

Explore conditioned generation in neural networks for natural language processing through this comprehensive lecture from Carnegie Mellon University's 2018 Neural Nets for NLP course. Delve into key concepts including language models, conditioned language models, and various generation techniques such as ancestral sampling and ensembling. Learn about evaluation methods like human evaluation and perplexity, and gain insights into the differences between linear and log-linear models. Discover advanced topics such as parameter averaging, ensemble distillation, and stacking. Conclude with a contrastive look at evaluating unconditioned generation, providing a well-rounded understanding of this crucial aspect of NLP.

Syllabus

Intro
Language Models Language models are generative models of text
Conditioned Language Models
Ancestral Sampling
Ensembling
Linear or Log Linear?
Parameter Averaging
Ensemble Distillation (e.g. Kim et al. 2016)
Stacking
Basic Evaluation Paradigm
Human Evaluation
Perplexity
A Contrastive Note: Evaluating Unconditioned Generation


Taught by

Graham Neubig

Related Courses

Microsoft Bot Framework and Conversation as a Platform
Microsoft via edX
Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube
Improving Customer Experiences with Speech to Text and Text to Speech
Microsoft via YouTube
Stanford Seminar - Deep Learning in Speech Recognition
Stanford University via YouTube
Select Topics in Python: Natural Language Processing
Codio via Coursera