The Key Equation Behind Probability - Entropy, Cross-Entropy, and KL Divergence
Offered By: Artem Kirsanov via YouTube
Course Description
Overview
Explore the fundamental concepts of probability theory and its applications in neuroscience and machine learning in this 26-minute video. Delve into the intuitive idea of surprise and its relation to probability through real-world examples. Examine advanced topics such as entropy, cross-entropy, and Kullback-Leibler (KL) divergence. Learn how to measure the average surprise in a probability distribution, understand the loss of information when approximating distributions, and quantify differences between probability distributions. Gain insights into Bayesian and Frequentist approaches to probability, probability distributions, and the role of objective functions in cross-entropy minimization.
Syllabus
Introduction
Sponsor: NordVPN
What is probability Bayesian vs Frequentist
Probability Distributions
Entropy as average surprisal
Cross-Entropy and Internal models
Kullback–Leibler KL divergence
Objective functions and Cross-Entropy minimization
Conclusion & Outro
Taught by
Artem Kirsanov
Related Courses
Introduction to Probability, Statistics, and Random ProcessesUniversity of Massachusetts Amherst via Independent Bayesian Statistics
Duke University via Coursera Bayesian Statistics: From Concept to Data Analysis
University of California, Santa Cruz via Coursera Improving your statistical inferences
Eindhoven University of Technology via Coursera Bayesian Statistics: Techniques and Models
University of California, Santa Cruz via Coursera