Demo: Approximate Bayesian Inference
Compare an exact grid-evaluated beta posterior with a local normal approximation. The goal is to see both when Laplace is useful and when its assumptions are fragile.
Mathematical setup
For beta-Bernoulli updating, the posterior density is proportional to
\[p(\theta\mid x)\propto \theta^{\alpha+s-1}(1-\theta)^{\beta+n-s-1}, \qquad 0<\theta<1.\]Laplace approximation expands the log posterior around an interior mode $\hat\theta$ where the negative second derivative is positive:
\[\log p(\theta\mid x) \approx \log p(\hat\theta\mid x) -\frac{1}{2}J(\hat\theta)(\theta-\hat\theta)^2,\]where $J(\hat\theta)=-\ell’’(\hat\theta)>0$. The resulting approximation is the normal density $N(\hat\theta,J(\hat\theta)^{-1})$. For this beta posterior with $\alpha+s>1$ and $\beta+n-s>1$,
\[\hat\theta=\frac{\alpha+s-1}{\alpha+\beta+n-2}.\]The red curve is a Laplace approximation only when this mode is interior. Otherwise the page labels it as a moment-matched normal fallback, which is a separate approximation and not an ordinary boundary-corrected Laplace approximation.
What to try
- Use moderate counts away from 0 and $n$. The Laplace curve should track the grid posterior well near the mode.
- Try $s=0$ or $s=n$ with weak priors. The boundary behavior makes the ordinary interior Laplace approximation invalid.
- Increase $n$. The posterior becomes more concentrated, and the local quadratic approximation usually improves when the mode remains interior.
The Laplace curve uses local second-order information at an interior mode. If the posterior mode moves to 0 or 1, the demo labels the red curve as a moment-matched normal fallback rather than a valid interior Laplace approximation.
