Discrepancy Minimization via Regularization
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore a groundbreaking algorithmic framework for discrepancy minimization based on regularization in this 57-minute lecture by Adrian Vladu from IRIF. Discover how varying the regularizer provides new interpretations of landmark works in algorithmic discrepancy, from Spencer's theorem to Banaszczyk's bound. Learn about the application of these techniques to prove the Beck-Fiala and Komlos conjectures for a novel regime of pseudorandom instances. Delve into the joint work with Lucas Pesenti, presented at SODA 2023, as part of the Optimization and Algorithm Design series at the Simons Institute.
Syllabus
Discrepancy Minimization via Regularization
Taught by
Simons Institute
Related Courses
Data Analysis and VisualizationGeorgia Institute of Technology via Udacity Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization
DeepLearning.AI via Coursera 機器學習基石下 (Machine Learning Foundations)---Algorithmic Foundations
National Taiwan University via Coursera Data Science: Machine Learning
Harvard University via edX Art and Science of Machine Learning auf Deutsch
Google Cloud via Coursera