Functional Gradient Descent Methods for Optimization and Sampling
Offered By: VinAI via YouTube
Course Description
Overview
Explore advanced machine learning concepts in this seminar on Functional Gradient Descent methods for optimization and sampling. Delve into the research of Dr. Qiang Liu, Assistant Professor of Computer Science at the University of Texas at Austin, as he discusses innovative applications of gradient descent beyond traditional Euclidean spaces. Learn about Stein variational gradient descent and its use in drawing samples from complex distributions. Discover how splitting steepest neural architecture descent can be applied to jointly estimate neural network weights and structures. Examine extensions of these techniques for handling constrained, bilevel, and multi-objective optimization problems. Gain insights into cutting-edge approaches that push the boundaries of statistical learning methods for high-dimensional and complex data.
Syllabus
Seminar Series: Functional Gradient Descent methods for optimization and sampling
Taught by
VinAI
Related Courses
Machine Learning Modeling Pipelines in ProductionDeepLearning.AI via Coursera MLOps for Scaling TinyML
Harvard University via edX Parameter Prediction for Unseen Deep Architectures - With First Author Boris Knyazev
Yannic Kilcher via YouTube SpineNet - Learning Scale-Permuted Backbone for Recognition and Localization
Yannic Kilcher via YouTube Synthetic Petri Dish - A Novel Surrogate Model for Rapid Architecture Search
Yannic Kilcher via YouTube