Training a Neural Network - Implementing Backpropagation and Gradient Descent from Scratch
Offered By: Valerio Velardo - The Sound of AI via YouTube
Course Description
Overview
Dive into a comprehensive video tutorial on implementing backpropagation and gradient descent from scratch using Python. Learn how to train a neural network to perform arithmetic sum operations. Explore key concepts including data representation, derivatives, reshaping, and the creation of a Natural Language Processing (NLP) model. Follow along as the instructor demonstrates the implementation of backpropagation, testing procedures, and the application of gradient descent. Gain hands-on experience in training a Multilayer Perceptron (MLP) and understand the intricacies of neural network training. Access the accompanying code on GitHub for further practice and experimentation.
Syllabus
Introduction
Data Representation
Derivatives
Reshape
Back propagation
Creating an NLP
Implementing backpropagation
Testing backpropagation
Implementing gradient descent
Applying gradient descent
Printing weights
Testing
Gradient Descent
Train
Train MLP
Taught by
Valerio Velardo - The Sound of AI
Related Courses
Neural Networks for Machine LearningUniversity of Toronto via Coursera 機器學習技法 (Machine Learning Techniques)
National Taiwan University via Coursera Machine Learning Capstone: An Intelligent Application with Deep Learning
University of Washington via Coursera Прикладные задачи анализа данных
Moscow Institute of Physics and Technology via Coursera Leading Ambitious Teaching and Learning
Microsoft via edX