Too Sharp, Too Sure: When Calibration Follows Curvature
Alessandro Morosini, Matea Gjika, Tomaso Poggio, Pierfrancesco Beneventano
TLDR
This paper shows neural network calibration is tightly linked to training-time curvature and introduces a margin-aware objective for better calibration.
Key contributions
- Discovers a tight coupling between neural network calibration, curvature, and margins during training.
- Shows Expected Calibration Error (ECE) closely tracks curvature-based sharpness throughout optimization.
- Mathematically proves ECE and Gauss-Newton curvature are controlled by the same margin-dependent functional.
- Introduces a margin-aware training objective improving out-of-sample calibration without accuracy loss.
Why it matters
Calibration is crucial for reliable AI systems, especially in high-stakes applications. This work provides a fundamental understanding of calibration during training, moving beyond post-hoc fixes. By offering a novel training objective, it enables the development of more trustworthy and robust neural networks.
Original Abstract
Modern neural networks can achieve high accuracy while remaining poorly calibrated, producing confidence estimates that do not match empirical correctness. Yet calibration is often treated as a post-hoc attribute. We take a different perspective: we study calibration as a training-time phenomenon on small vision tasks, and ask whether calibrated solutions can be obtained reliably by intervening on the training procedure. We identify a tight coupling between calibration, curvature, and margins during training of deep networks under multiple gradient-based methods. Empirically, Expected Calibration Error (ECE) closely tracks curvature-based sharpness throughout optimization. Mathematically, we show that both ECE and Gauss--Newton curvature are controlled, up to problem-specific constants, by the same margin-dependent exponential tail functional along the trajectory. Guided by this mechanism, we introduce a margin-aware training objective that explicitly targets robust-margin tails and local smoothness, yielding improved out-of-sample calibration across optimizers without sacrificing accuracy.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.