YoVDO

Neural Nets for NLP 2017 - Convolutional Networks for Text

Offered By: Graham Neubig via YouTube

Tags

Neural Networks Courses Natural Language Processing (NLP) Courses Convolution Courses Bag of Words Courses

Course Description

Overview

Explore convolutional networks for text processing in this comprehensive lecture from CMU's Neural Networks for NLP course. Dive into bag-of-words models, n-grams, and convolution techniques. Learn about context windows, sentence modeling, and advanced concepts like stacked and dilated convolutions. Discover structured convolution methods and approaches for modeling sentence pairs. Gain insights into visualizing convolutional neural networks for text analysis. Access accompanying slides and code examples to reinforce your understanding of these powerful techniques in natural language processing.

Syllabus

Intro
An Example Prediction Problem: Sentence Classification
A First Try: Bag of Words (BOW)
Continuous Bag of Words (CBOW) movie
What do Our Vectors Represent?
Why Bag of n-grams?
What Problems w/ Bag of n-grams?
Time Delay Neural Networks (Waibel et al. 1989)
Convolutional Networks (LeCun et al. 1997)
Standard conv2d Function
Stacked Convolution
Dilated Convolution (e.g. Kalchbrenner et al. 2016)
An Aside: Nonlinear Functions • Proper choice of a non-linear function is essential in stacked networks
Why (Dilated) Convolution for Modeling Sentences? • In contrast to recurrent neural networks (next class)
Example: Dependency Structure
Why Model Sentence Pairs?
Siamese Network (Bromley et al. 1993)


Taught by

Graham Neubig

Related Courses

Neural Networks for Machine Learning
University of Toronto via Coursera
Good Brain, Bad Brain: Basics
University of Birmingham via FutureLearn
Statistical Learning with R
Stanford University via edX
Machine Learning 1—Supervised Learning
Brown University via Udacity
Fundamentals of Neuroscience, Part 2: Neurons and Networks
Harvard University via edX