ArXiv TLDR

Socrates Loss: Unifying Confidence Calibration and Classification by Leveraging the Unknown

🐦 Tweet
2604.12245

Sandra Gómez-Gálvez, Tobias Olenyi, Gillian Dobbie, Katerina Taškova

cs.LGcs.AIcs.CVcs.NE

TLDR

Socrates Loss unifies confidence calibration and classification in deep neural networks by introducing an auxiliary "unknown" class and dynamic uncertainty penalty.

Key contributions

  • Proposes Socrates Loss, a unified function for simultaneous classification and confidence calibration.
  • Leverages uncertainty via an auxiliary "unknown" class and a dynamic uncertainty penalty.
  • Achieves improved training stability and a better accuracy-calibration trade-off.
  • Provides theoretical guarantees to prevent miscalibration and overfitting.

Why it matters

Deep neural networks often lack reliable confidence calibration, hindering their use in critical applications. Socrates Loss unifies classification and calibration, overcoming the trade-off in existing methods. This results in more stable and accurate models, enhancing reliability for high-stakes scenarios.

Original Abstract

Deep neural networks, despite their high accuracy, often exhibit poor confidence calibration, limiting their reliability in high-stakes applications. Current ad-hoc confidence calibration methods attempt to fix this during training but face a fundamental trade-off: two-phase training methods achieve strong classification performance at the cost of training instability and poorer confidence calibration, while single-loss methods are stable but underperform in classification. This paper addresses and mitigates this stability-performance trade-off. We propose Socrates Loss, a novel, unified loss function that explicitly leverages uncertainty by incorporating an auxiliary unknown class, whose predictions directly influence the loss function and a dynamic uncertainty penalty. This unified objective allows the model to be optimized for both classification and confidence calibration simultaneously, without the instability of complex, scheduled losses. We provide theoretical guarantees that our method regularizes the model to prevent miscalibration and overfitting. Across four benchmark datasets and multiple architectures, our comprehensive experiments demonstrate that Socrates Loss consistently improves training stability while achieving more favorable accuracy-calibration trade-off, often converging faster than existing methods.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.