Automated Scalable Bayesian Inference via Data Summarization - 2018
Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube
Course Description
Overview
Explore advanced techniques for scalable Bayesian inference in large-scale data settings through this 59-minute lecture by MIT's Tamara Broderick. Delve into the concept of data summarization and coresets as a means to overcome computational challenges in Bayesian methods. Learn about theoretical guarantees on coreset size and approximation quality, and discover how this approach provides geometric decay in posterior approximation error. Examine the application of these techniques to both synthetic and real datasets, demonstrating significant improvements over uniform random subsampling. Gain insights into Broderick's research on developing and analyzing models for scalable Bayesian machine learning, and understand the potential impact of these methods on handling large datasets efficiently while maintaining the benefits of Bayesian inference.
Syllabus
Core" of the data set • Observe: redundancies can exist even if data isn't "tall
Roadmap
Bayesian coresets
Uniform subsampling revisited
Importance sampling
Hilbert coresets
Frank-Wolfe
Gaussian model (simulated) • 1K pts; norms, inference: closed-form
Logistic regression (simulated)
Real data experiments
Data summarization
Taught by
Center for Language & Speech Processing(CLSP), JHU
Related Courses
Infinite Memory Transformer - Research Paper ExplainedYannic Kilcher via YouTube Rendering Games with Millions of Ray Traced Lights
Nvidia via YouTube Brooklyn Quant Experience Lecture Series - Fourier-Based Methods for Complex Insurance Products Management
New York University (NYU) via YouTube Batch Offline Reinforcement Learning - Part 1
Simons Institute via YouTube A Classical Algorithm Framework for Dequantizing Quantum Machine Learning
Simons Institute via YouTube