YoVDO

Bayesian Inference with MCMC

Offered By: Databricks via Coursera

Tags

Bayesian Statistics Courses Statistics & Probability Courses Python Courses

Course Description

Overview

The objective of this course is to introduce Markov Chain Monte Carlo Methods for Bayesian modeling and inference, The attendees will start off by learning the the basics of Monte Carlo methods. This will be augmented by hands-on examples in Python that will be used to illustrate how these algorithms work. This will be the second course in a specialization of three courses .Python and Jupyter notebooks will be used throughout this course to illustrate and perform Bayesian modeling with PyMC3. The course website is located at https://sjster.github.io/introduction_to_computational_statistics/docs/index.html. The course notebooks can be downloaded from this website by following the instructions on page https://sjster.github.io/introduction_to_computational_statistics/docs/getting_started.html. The instructor for this course will be Dr. Srijith Rajamohan.

Syllabus

  • Topics in Model Performance
    • This module gives an overview of topics related to assessing the quality of models. While some of these metrics may be familiar to those with a Machine Learning background, the goal is to bring awareness to the concepts rooted in Information Theory. The course website is https://sjster.github.io/introduction_to_computational_statistics/docs/Production/BayesianInference.html. Instructions to download and run the notebooks are at https://sjster.github.io/introduction_to_computational_statistics/docs/Production/getting_started.html
  • The Metropolis Algorithms for MCMC
    • This module serves as a gentle introduction to Markov-Chain Monte Carlo methods. The general idea behind Markov chains are presented along with their role in sampling from distributions. The Metropolis and Metropolis-Hastings algorithms are introduced and implemented in Python to help illustrate their details. The course website is https://sjster.github.io/introduction_to_computational_statistics/docs/Production/MonteCarlo.html. Instructions to download and run the notebooks are at https://sjster.github.io/introduction_to_computational_statistics/docs/Production/getting_started.html
  • Gibbs Sampling and Hamiltonian Monte Carlo Algorithms
    • This module is a continuation of module 2 and introduces Gibbs sampling and the Hamiltonian Monte Carlo (HMC) algorithms for inferring distributions. The Gibbs sampler algorithm is illustrated in detail, while the HMC receives a more high-level treatment due to the complexity of the algorithm. Finally, some of the properties of MCMC algorithms are presented to set the stage for Course 3 which uses the popular probabilistic framework PyMC3. The course website is https://sjster.github.io/introduction_to_computational_statistics/docs/Production/MonteCarlo.html#gibbs-sampling. Instructions to download and run the notebooks are at https://sjster.github.io/introduction_to_computational_statistics/docs/Production/getting_started.html

Taught by

Dr. Srijith Rajamohan

Related Courses

Accounting for Death in War: Separating Fact from Fiction
Royal Holloway, University of London via FutureLearn
Advanced Machine Learning
The Open University via FutureLearn
Advanced Statistics for Data Science
Johns Hopkins University via Coursera
農企業管理學 (Agribusiness Management)
National Taiwan University via Coursera
AI & Machine Learning
Arizona State University via Coursera