YoVDO

Math for Machine Learning - Exercises: Probability

Offered By: Weights & Biases via YouTube

Tags

Statistics & Probability Courses Machine Learning Courses Gradient Descent Courses Probability Courses Entropy Courses Loss Functions Courses Gaussian Distribution Courses Divergence Courses

Course Description

Overview

Dive into a 44-minute video tutorial on probability exercises for machine learning, led by Weights & Biases experts Charles Frye and Scott Condron. Work through practical implementations of entropy, surprise, cross-entropy, and divergence concepts. Explore the relationship between loss functions and surprises, and apply these principles to Gaussian distributions. Gain hands-on experience with gradient descent on surprise functions. Understand the importance of probability in machine learning models and discover how programmers can enhance their mathematical skills for ML applications. Access additional resources, including GitHub exercises and related lectures, to further deepen your understanding of probability in the context of machine learning.

Syllabus

- Teaser
- Intro
- Implementing entropies
- Exercise: surprise
- Why do models output probabilities?
- Exercise: entropy
- Exercise: crossentropy
- Exercise: divergence
- Loss functions and surprises
- Exercise: softmax_crossentropy
- Putting it all together with Gaussians
- Exercise: gaussian_surprise
- Gaussian surprise and squared error
- Exercise: Gradient descent on a surprise
- Why are these exercises useful?
- How programmers can learn more math


Taught by

Weights & Biases

Related Courses

AI skills for Engineers: Supervised Machine Learning
Delft University of Technology via edX
Audio Classification with TensorFlow
Coursera Project Network via Coursera
Deep Learning - Artificial Neural Networks with TensorFlow
Packt via Coursera
Deep Neural Network for Beginners Using Python
Packt via Coursera
Introduction to RNN and DNN
Packt via Coursera