YoVDO

Remaining Challenges in Deep Learning Based NLP

Offered By: WeAreDevelopers via YouTube

Tags

WeAreDevelopers World Congress Courses Big Data Courses Deep Learning Courses Overfitting Courses Representation Learning Courses Interpretability Courses Sentence Embedding Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the remaining challenges in Deep Learning-based Natural Language Processing (NLP) in this insightful 31-minute conference talk. Delve into the limitations of current neural models, including their reliance on large training datasets, potential biases in performance metrics, and lack of robustness. Examine the shortcomings of distributional word embeddings and sentence-level representations. Investigate the implications of limited interpretability in neural networks, affecting both debugging processes and fairness issues. Gain a critical perspective on the state of AI in language processing and understand the areas that still require significant improvement in the field of NLP.

Syllabus

Introduction
About me
Popular media examples
Adding more layers
Representation learning
Recurrent neural networks
Big data problem
Overfitting
Life cycle
Distributional similarity
Sentence embedding
Not robust enough
Lack of interpretability
Questions


Taught by

WeAreDevelopers

Related Courses

From Graph to Knowledge Graph – Algorithms and Applications
Microsoft via edX
Social Network Analysis
Indraprastha Institute of Information Technology Delhi via Swayam
Stanford Seminar - Representation Learning for Autonomous Robots, Anima Anandkumar
Stanford University via YouTube
Unsupervised Brain Models - How Does Deep Learning Inform Neuroscience?
Yannic Kilcher via YouTube
Emerging Properties in Self-Supervised Vision Transformers - Facebook AI Research Explained
Yannic Kilcher via YouTube