ArXiv TLDR

Hyperbolic Concept Bottleneck Models

🐦 Tweet
2605.06440

Daniel Uyterlinde, Swasti Shreya Mishra, Pascal Mettes

cs.LGcs.CV

TLDR

Hyperbolic Concept Bottleneck Models (HypCBM) improve interpretability by embedding concepts in hyperbolic space, leveraging semantic hierarchies.

Key contributions

  • Introduces HypCBM, grounding concept bottlenecks in hyperbolic space to reflect semantic hierarchies.
  • Reformulates concept activation as asymmetric geometric containment, yielding sparse, hierarchy-aware signals.
  • Proposes an adaptive scaling law for coherent, hierarchically faithful interventions and corrections.
  • Achieves strong interpretability, consistency, and robustness, outperforming Euclidean models in sparse data.

Why it matters

Current Concept Bottleneck Models treat concepts as independent, ignoring their inherent hierarchical structure. This paper resolves this by using hyperbolic geometry, leading to more accurate and robust concept activations. It significantly improves interpretability and consistency in AI models.

Original Abstract

Concept Bottleneck Models (CBMs) have become a popular approach to enable interpretability in neural networks by constraining classifier inputs to a set of human-understandable concepts. While effective, current models embed concepts in flat Euclidean space, treating them as independent, orthogonal dimensions. Concepts, however, are highly structured and organized in semantic hierarchies. To resolve this mismatch, we propose Hyperbolic Concept Bottleneck Models (HypCBM), a post-hoc framework that grounds the bottleneck in this structure by reformulating concept activation as asymmetric geometric containment in hyperbolic space. Rather than treating entailment cones as a pre-training penalty, we show they encode a natural test-time activation signal: the margin of inclusion within a concept's entailment cone yields sparse, hierarchy-aware activations without any additional supervision or learned modules. We further introduce an adaptive scaling law for hierarchically faithful interventions, propagating user corrections coherently through the concept tree. Empirically, HypCBM rivals post-hoc Euclidean models trained on 20$\times$ more data in sparse regimes required for human interpretability, with stronger hierarchical consistency and improved robustness to input corruptions.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.