Bayesian Methods

March 2026
Aathreya Kadambi

The Bayesian Update

The main idea behind Bayesian inference is Bayes’ rule:

p(yx)=p(xy)p(y)p(x)p(xy)p(y)p(y \mid x) = \dfrac{p(x \mid y)p(y)}{p(x)} \propto p(x \mid y)p(y)
The symbols are interpretted in the following way:

  • xx: observed data,
  • yy: modeling parameters/the underlying universe state/something else depending on your philosophy,
  • p(yx)p(y\mid x): the posterior distribution of universe states given observation of data xx,
  • p(y)p(y): a prior belief about the distribution of universe states before observing xx,
  • p(xy)p(x \mid y): the likelihood of obesrving some data xx given the universe state being yy,
  • p(x)p(x): the evidence for data xx.

The evidence term is usually computed by integrating over the joint distribution p(x,y)p(x, y), which is computationally expensive. As such, we often prefer the form:

logp(yx)=+Clogp(xy)+logp(y)\log p(y \mid x) \overset{+C}{=} \log p(x \mid y) + \log p(y)
In other words, we can specify the log posterior up to a constant, if we know the log likelihood and log prior.

According to Andrew Gelman’s book (and Professor Alexander Strang whose course I took), there are three important pieces to Bayesian statistics:

  1. Model Specification: This is a specification of what distributions will be used for the prior and likelihood.
  2. Bayesian Update: This is computing or approximating the posterior given by the Bayesian update.
  3. Model Evaluation: This is the evaluation of the model and interpretation of the results.

Each of these broad questions sits on top of large bodies of research, analysis, and thought. I hope to explore some of that in these notes.

In my opinion, Bayesian statistics is one of the most philosophically pleasing subjects one can study in life because it seems to give a very casual and stable interpretation of the absence of meaning or objective truth. It also explains several phenomena that people might initially view as signs of inexplicable complexity. It is also very flexible, allowing the user to impose their own ideas or truths while also giving enough room for the data to tell its own story.

Large Sample Theory and Convergence: The Big Data Case [Model Sepcficiation]

Vibe. In the big data case (when we have a lot of data xx), the power of Bayesian inference comes from convergence: the modeling paramters will converge to the parameters which minimize some notion of distributional distance to true phenomenon underlying the data.

We will make this precise with several theorems below.

Importance of Prior: The Small Data Case [Model Specification]

Vibe. In the small data case (when we don’t have much data xx), the power of Bayesian inference comes from the power of the user to specify an informative prior. A prior can be thought of as a model fit on virtual data points, with stronger priors corresponding to more virtual data.

We will again make this precise with several theorems below.

Conjugate Priors Are Invariant to Bayesian Updates [Model Specification]

Beta-Binomial

Weighted Update Rules Can Also Make Sense [Bayesian Update]



As a fun fact, it might seem like this website is flat because you're viewing it on a flat screen, but the curvature of this website actually isn't zero. ;-)

Copyright © 2026, Aathreya Kadambi

Made with Astrojs, React, and Tailwind.