ArXiv TLDR

Concentration and Calibration in Predictive Bayesian Inference

🐦 Tweet
2605.00455

David T. Frazier, Hui Wang

stat.MEmath.STstat.ML

TLDR

Predictive Bayesian Inference's reliability and calibration depend critically on the accuracy of its forward predictive model, which can lead to poor coverage.

Key contributions

  • PBI posteriors concentrate on a quantity explicitly tied to the forward predictive model used.
  • The forward predictive model entirely dictates the uncertainty quantification produced by PBI.
  • Inaccurate forward predictive models can lead to PBI credible sets with near-zero coverage.
  • Calibrated PBI inferences require the predictive engine to accurately capture the true data generating process.

Why it matters

This paper clarifies the critical role of the forward predictive model in Predictive Bayesian Inference (PBI). It demonstrates that PBI's calibration and reliability hinge entirely on this model's accuracy, providing crucial guidance for practitioners.

Original Abstract

Predictive Bayesian inference (PBI) represents a model-and prior-agnostic approach to standard Bayesian inference which allows users to quantify uncertainty for a functional of interest only by specifying a forward predictive model for future unobserved data. The flexibility and generality of this framework have led to a host of novel algorithms for implementing this approach, and many empirical applications, yet the reliability of the resulting inferences for the underlying statistical functional of interest remains unclear. Herein, we demonstrate that when using PBI for a population functional of interest, the resulting posterior concentrates onto a well-defined quantity that explicitly depends on the forward predictive model used to implement the predictive recursion underlying the method. Furthermore, the forward predictive model entirely determines the uncertainty quantification produced in PBI. Consequently, our results show that if the predictive model does not capture all relevant features of the data, and, even in very simple examples, the coverage of predictive Bayes credible sets for the population value of the functional of interest can be arbitrarily close to zero. We carefully explain why this occurs, and show that this behavior is directly tied to the inaccuracy of the forward predictive model used to produce future observations within the PBI framework. As a consequence, our results imply that in order for PBI to deliver calibrated posterior inferences, the resulting predictive engine used to generate posterior samples must contain, in a well-defined sense, the true DGP, else inferences generated under this framework will not be calibrated.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.