CMU Neural Nets for NLP - Distributional Semantics and Word Vectors
Offered By: Graham Neubig via YouTube
Course Description
Overview
Explore distributional semantics and word vectors in this comprehensive lecture from CMU's Neural Networks for NLP course. Delve into describing words by their context, counting and predicting techniques, skip-grams and CBOW models, and methods for evaluating and visualizing word vectors. Learn about advanced word vector techniques, including contextualization, WordNet, and various embedding models like CBOW and GloVe. Examine different types of context, evaluation methods for embeddings, and their practical applications in NLP systems. Discuss the limitations of embeddings and explore sub-word embedding techniques. Gain valuable insights into the foundations of word representations in natural language processing.
Syllabus
Remember: Neural Models
How to Train Embeddings?
What do we want to know about words?
Contextualization of Word Representations
A Manual Attempt: WordNet
An Answer (?): Word Embeddings!
Distributional vs. Distributed Representations
Distributional Representations (see Goldberg 10.4.1)
Word Embeddings from Language Models
Context Window Methods
CBOW (Mikolov et al. 2013) • Predict word based on sum of surrounding embeddings
Glove (Pennington et al. 2014)
What Contexts?
Types of Evaluation
Intrinsic Evaluation of Embeddings (categorization from Schnabel et al 2015)
Extrinsic Evaluation: Using Word Embeddings in Systems
How Do I Choose Embeddings?
Limitations of Embeddings
Sub-word Embeddings (2)
Taught by
Graham Neubig
Related Courses
Natural Language Processing with spaCy & Python - Course for BeginnersfreeCodeCamp Complete Natural Language Processing Tutorial in Python
Keith Galli via YouTube Introduction to NLP and Topic Modeling
Open Data Science via YouTube Understanding Unstructured Data with Language Models - Alex Peattie
Open Data Science via YouTube Experimental Creative Writing with the Vectorized Word
Strange Loop Conference via YouTube