ArXiv TLDR

Bayesian experimental design: grouped geometric pooled posterior via ensemble Kalman methods

🐦 Tweet
2604.18505

Huchen Yang, Xinghao Dong, Jinlong Wu

cs.ITstat.ML

TLDR

This paper introduces a grouped geometric pooled posterior framework using tailored Ensemble Kalman methods to improve Bayesian experimental design accuracy and efficiency.

Key contributions

  • Introduces a grouped geometric pooled posterior framework for Bayesian experimental design.
  • Uses tailored Ensemble Kalman Inversion (EKI) to generate group-specific proposals efficiently.
  • Develops a conservative diagnostic to guide grouping and assess importance-sampling quality.
  • Achieves more accurate and stable EIG estimation at costs comparable to amortized methods.

Why it matters

Bayesian experimental design for complex systems struggles with accuracy-efficiency trade-offs. This paper's grouped posterior framework improves expected information gain estimation accuracy and stability at comparable costs, enabling reliable design.

Original Abstract

Bayesian experimental design (BED) for complex physical systems is often limited by the nested inference required to estimate the expected information gain (EIG) or its gradients. Each outer sample induces a different posterior, creating a large and heterogeneous set of inference targets. Existing methods have to sacrifice either accuracy or efficiency: they either perform per-outer-sample posterior inference, which yields higher fidelity but at prohibitive computational cost, or amortize the inner inference across all outer samples for computational reuse, at the risk of degraded accuracy under posterior heterogeneity. To improve accuracy and maintain cost at the amortized level, we propose a grouped geometric pooled posterior framework that partitions outer samples into groups and constructs a pooled proposal for each group. While such grouping strategy would normally require generating separate proposal samples for different groups, our tailored ensemble Kalman inversion (EKI) formulation generates these samples without extra forward-model evaluation cost. We also introduce a conservative diagnostic to assess importance-sampling quality to guide grouping. This grouping strategy improves within-group proposal-target alignment, yielding more accurate and stable estimators while keeping the cost comparable to amortized approaches. We evaluate the performance of our method on both Gaussian-linear and high-dimensional network-based model discrepancy calibration problems.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.