Optimal Transport for Machine Learning - Gabriel Peyre, Ecole Normale Superieure
Offered By: Alan Turing Institute via YouTube
Course Description
Overview
Explore optimal transport techniques for machine learning in this 42-minute conference talk by Gabriel Peyre from Ecole Normale Superieure. Delve into probability distributions in data sciences, focusing on Kantorovitch's formulation, optimal transport distances, and entropic regularization. Examine Sinkhorn divergences and sample complexity before transitioning to density fitting and generative models. Compare deep discriminative and generative models, discussing training architectures and automatic differentiation. Analyze examples of image generation and generative adversarial networks. Conclude by considering open problems in the field. This talk, part of the Isaac Newton Institute's workshop on "Approximation, sampling and compression in data science," bridges various mathematical aspects of data science, fostering collaboration among researchers in computational statistics, machine learning, optimization, information theory, and learning theory.
Syllabus
Intro
Probability Distributions in Data Sciences
1. Optimal Transport
Kantorovitch's Formulation
Optimal Transport Distances
Entropic Regularization
Sinkhorn Divergences
Sample Complexity
Density Fitting and Generative Models
Deep Discriminative vs Generative Models
Training Architecture
Automatic Differentiation
Examples of Images Generation
Generative Adversarial Networks
Open Problems
Taught by
Alan Turing Institute
Related Courses
Beyond Worst-Case Analysis - Panel DiscussionSimons Institute via YouTube Reinforcement Learning - Part I
Simons Institute via YouTube Reinforcement Learning in Feature Space: Complexity and Regret
Simons Institute via YouTube Exploration with Limited Memory - Streaming Algorithms for Coin Tossing, Noisy Comparisons, and Multi-Armed Bandits
Association for Computing Machinery (ACM) via YouTube Learning Quantum with Generative Models
APS Physics via YouTube