Subtractive Mixture Models: Representation and Learning
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore a 52-minute lecture on subtractive mixture models presented by Antonio Vergari from the University of Edinburgh as part of the Simons Institute's series on Probabilistic Circuits and Logic. Delve into the concept of subtractive mixtures, which can reduce the number of components needed to model complex distributions by allowing the subtraction of probability mass or density. Examine the challenges of learning these models while maintaining non-negative functions, and discover how deep subtractive mixtures can be learned and inferred by squaring them within the framework of probabilistic circuits. Understand how this approach enables the representation of tensorized mixtures and generalizes other subtractive models like positive semi-definite kernel models and Born machines. Learn about the theoretical proof demonstrating that squared circuits with subtractions can be exponentially more expressive than traditional additive mixtures. Analyze empirical evidence showcasing this increased expressiveness in real-world distribution estimation tasks, and discuss the tractable inference scenarios for this new class of circuits.
Syllabus
Subtractive Mixture Models: Representation and Learning
Taught by
Simons Institute
Related Courses
Neural Networks for Machine LearningUniversity of Toronto via Coursera 機器學習技法 (Machine Learning Techniques)
National Taiwan University via Coursera Machine Learning Capstone: An Intelligent Application with Deep Learning
University of Washington via Coursera Прикладные задачи анализа данных
Moscow Institute of Physics and Technology via Coursera Leading Ambitious Teaching and Learning
Microsoft via edX