Neural Nets for NLP - Debugging Neural Nets for NLP
Offered By: Graham Neubig via YouTube
Course Description
Overview
Syllabus
Intro
In Neural Networks, Tuning is Paramount!
A Typical Situation
Identifying Training Time Problems
Is My Model Too Weak?
Be Careful of Deep Models
Trouble w/ Optimization
Reminder: Optimizers - SGD: take a step in the direction of the gradient
Learning Rate Learning rate is an important parameter
Initialization
Debugging Minibatching
Debugging Decoding
Debugging Search
Look At Your Data!
Quantitative Analysis
Symptoms of Overfitting
Reminder: Early Stopping, Learning Rate Decay
Reminder: Dropout (Srivastava et al. 2014) Neural nets have lots of parameters, and are prone to overfitting • Dropout: randomly zero-out nodes in the hidden layer with probability p at training time only
A Stark Example (Koehn and Knowles 2017) • Better search (=better model score) can result in worse BLEU score!
Managing Loss Function/ Eval Metric Differences Most principled way: use structured prediction techniques to be discussed in future classes
A Simple Method: Early Stopping w/ Eval Metric
Reproducing Previous Work
Taught by
Graham Neubig
Related Courses
Neural Networks for Machine LearningUniversity of Toronto via Coursera Good Brain, Bad Brain: Basics
University of Birmingham via FutureLearn Statistical Learning with R
Stanford University via edX Machine Learning 1—Supervised Learning
Brown University via Udacity Fundamentals of Neuroscience, Part 2: Neurons and Networks
Harvard University via edX