Neural Nets for NLP - Recurrent Networks for Sentence or Language Modeling
Offered By: Graham Neubig via YouTube
Course Description
Overview
Syllabus
Intro
Why Model Sentence Pairs?
Siamese Network (Bromley et al. 1993)
Convolutional Matching Model (Hu et al. 2014) • Concatenate sentences into a 30 tensor and perform convolution
Convolutional Features + Matrix-based Pooling in and Schutze 2015
NLP and Sequential Data
Long-distance Dependencies in Language
Can be Complicated!
Recurrent Neural Networks (Elman 1990)
Unrolling in Time • What does processing a sequence look like?
What Can RNNs Do?
Representing Sentences
e.g. Language Modeling
RNNLM Example: Loss Calculation and State Update
Vanishing Gradient • Gradients decrease as they get pushed back
LSTM Structure
What can LSTMs Learn? (2) (Shi et al. 2016, Radford et al. 2017) Count length of sentence
Handling Long Sequences
Taught by
Graham Neubig
Related Courses
Natural Language ProcessingColumbia University via Coursera Natural Language Processing
Stanford University via Coursera Introduction to Natural Language Processing
University of Michigan via Coursera moocTLH: Nuevos retos en las tecnologías del lenguaje humano
Universidad de Alicante via Miríadax Natural Language Processing
Indian Institute of Technology, Kharagpur via Swayam