YoVDO

Neural Nets for NLP 2017 - Convolutional Networks for Text

Offered By: Graham Neubig via YouTube

Tags

Neural Networks Courses Natural Language Processing (NLP) Courses Convolution Courses Bag of Words Courses

Course Description

Overview

Explore convolutional networks for text processing in this comprehensive lecture from CMU's Neural Networks for NLP course. Dive into bag-of-words models, n-grams, and convolution techniques. Learn about context windows, sentence modeling, and advanced concepts like stacked and dilated convolutions. Discover structured convolution methods and approaches for modeling sentence pairs. Gain insights into visualizing convolutional neural networks for text analysis. Access accompanying slides and code examples to reinforce your understanding of these powerful techniques in natural language processing.

Syllabus

Intro
An Example Prediction Problem: Sentence Classification
A First Try: Bag of Words (BOW)
Continuous Bag of Words (CBOW) movie
What do Our Vectors Represent?
Why Bag of n-grams?
What Problems w/ Bag of n-grams?
Time Delay Neural Networks (Waibel et al. 1989)
Convolutional Networks (LeCun et al. 1997)
Standard conv2d Function
Stacked Convolution
Dilated Convolution (e.g. Kalchbrenner et al. 2016)
An Aside: Nonlinear Functions • Proper choice of a non-linear function is essential in stacked networks
Why (Dilated) Convolution for Modeling Sentences? • In contrast to recurrent neural networks (next class)
Example: Dependency Structure
Why Model Sentence Pairs?
Siamese Network (Bromley et al. 1993)


Taught by

Graham Neubig

Related Courses

Computational Photography
Georgia Institute of Technology via Udacity
Discrete Time Signals and Systems, Part 1: Time Domain
Rice University via edX
Signals and Systems, Part 1
Indian Institute of Technology Bombay via edX
Discrete Time Signals and Systems, Part 2: Frequency Domain
Rice University via edX
Introduction to Sound and Acoustic Sketching
University St. Joseph via Kadenze