Week 1: Bayesian Methods, MCMC & Regularization Theory
Deep dive into Bayesian inference, Markov Chain Monte Carlo, regularization theory, and advanced model selection techniques for research-level ML.
- Derive the Bayesian posterior for common conjugate models
- Implement Markov Chain Monte Carlo (MCMC) from scratch
- Prove the Lasso and Ridge regularization properties
- Apply cross-validation theory for model selection
This first lecture establishes the foundational framework for Advanced Statistical Learning. By the end of this session, you will have the conceptual grounding and practical starting point needed for the rest of the course.
Key Concepts
The lecture introduces the four main pillars of this course: Bayesian Inference & Conjugate Priors, MCMC: Metropolis-Hastings & Gibbs Sampling, Regularization Theory: Lasso & Ridge, Information Criteria: AIC, BIC, WAIC. Each will be explored in depth over the 14-week curriculum, with hands-on projects reinforcing theory at every stage.
This Week's Focus
Focus on mastering: Bayesian Inference & Conjugate Priors and MCMC: Metropolis-Hastings & Gibbs Sampling. These are the prerequisites for everything in Week 2. The concepts build on each other — do not skip the practice exercises.
DS501 Project 1: Bayesian Regression with MCMC
Implement Bayesian linear regression with MCMC sampling using PyMC or Stan. Compare posterior predictive distributions against frequentist estimates on a regression dataset.
- MCMC sampler implementation (from scratch or PyMC)
- Convergence diagnostics (R-hat, effective sample size)
- Posterior predictive checks and calibration plots
- Comparison with MLE and LASSO estimates
These represent the style and difficulty of questions you'll see on the midterm and final. Start thinking about them now.
Derive the posterior distribution for a Gaussian likelihood with a Gaussian prior on the mean.
What is the Metropolis-Hastings algorithm? Prove that it converges to the target distribution.
When does Lasso produce exact zeros in coefficients? Prove using the KKT conditions.