Self-Supervised Laplace Approximation for Bayesian Uncertainty Quantification
Julian Rodemann, Alexander Marquard, Thomas Augustin, Michele Caprio
TLDR
Introduces Self-Supervised Laplace Approximation (SSLA) to quantify Bayesian model predictive uncertainty by refitting on self-predicted data, outperforming classical methods.
Key contributions
- Directly approximates the posterior predictive distribution, bypassing complex parameter posteriors.
- Quantifies predictive uncertainty by refitting Bayesian models on self-predicted data.
- Offers a deterministic, sampling-free, and computationally efficient method (SSLA/ASSLA).
- Outperforms classical Laplace approximations in predictive calibration across various tasks.
Why it matters
This paper simplifies Bayesian uncertainty quantification by directly approximating the posterior predictive, bypassing complex parameter posteriors. The proposed SSLA/ASSLA method is deterministic, sampling-free, and computationally efficient. It significantly improves predictive calibration over traditional Laplace approximations, enhancing the reliability of Bayesian models.
Original Abstract
Approximate Bayesian inference typically revolves around computing the posterior parameter distribution. In practice, however, the main object of interest is often a model's predictions rather than its parameters. In this work, we propose to bypass the parameter posterior and focus directly on approximating the posterior predictive distribution. We achieve this by drawing inspiration from self-training within self-supervised and semi-supervised learning. Essentially, we quantify a Bayesian model's predictive uncertainty by refitting on self-predicted data. The idea is strikingly simple: If a model assigns high likelihood to self-predicted data, these predictions are of low uncertainty, and vice versa. This yields a deterministic, sampling-free approximation of the posterior predictive. The modular structure of our Self-Supervised Laplace Approximation (SSLA) further allows us to plug in different prior specifications, enabling classical Bayesian sensitivity (w.r.t. prior choice) analysis. In order to bypass expensive refitting, we further introduce an approximate version of SSLA, called ASSLA. We study (A)SSLA both theoretically and empirically in regression models ranging from Bayesian linear models to Bayesian neural networks. Across a wide array of regression tasks with simulated and real-world datasets, our methods outperform classical Laplace approximations in predictive calibration while remaining computationally efficient.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.