Neural Nets for NLP - Recurrent Networks for Sentence or Language Modeling
Offered By: Graham Neubig via YouTube
Course Description
Overview
Syllabus
Intro
Why Model Sentence Pairs?
Siamese Network (Bromley et al. 1993)
Convolutional Matching Model (Hu et al. 2014) • Concatenate sentences into a 30 tensor and perform convolution
Convolutional Features + Matrix-based Pooling in and Schutze 2015
NLP and Sequential Data
Long-distance Dependencies in Language
Can be Complicated!
Recurrent Neural Networks (Elman 1990)
Unrolling in Time • What does processing a sequence look like?
What Can RNNs Do?
Representing Sentences
e.g. Language Modeling
RNNLM Example: Loss Calculation and State Update
Vanishing Gradient • Gradients decrease as they get pushed back
LSTM Structure
What can LSTMs Learn? (2) (Shi et al. 2016, Radford et al. 2017) Count length of sentence
Handling Long Sequences
Taught by
Graham Neubig
Related Courses
Reinforcement Learning for Trading StrategiesNew York Institute of Finance via Coursera Natural Language Processing with Sequence Models
DeepLearning.AI via Coursera Fake News Detection with Machine Learning
Coursera Project Network via Coursera English/French Translator: Long Short Term Memory Networks
Coursera Project Network via Coursera Text Classification Using Word2Vec and LSTM on Keras
Coursera Project Network via Coursera