Quantifying Explanation Consistency: The C-Score Metric for CAM-Based Explainability in Medical Image Classification
Kabilan Elangovan, Daniel Ting
TLDR
This paper introduces C-Score, a novel metric to quantify explanation consistency for CAM methods in medical imaging, revealing model instability before AUC collapse.
Key contributions
- Introduces C-Score, an annotation-free metric to quantify intra-class explanation reproducibility for CAMs.
- Evaluates six CAM techniques across three CNN architectures on the Kermany chest X-ray dataset.
- Identifies three distinct mechanisms of AUC-consistency dissociation, invisible to standard metrics.
- C-Score provides an early warning signal of impending model instability, detecting deterioration before AUC collapse.
Why it matters
This paper introduces a crucial metric for evaluating the reliability of AI explanations in critical domains like medical imaging. By quantifying consistency, it helps identify unstable models early, improving trust and safety. It shifts focus beyond predictive accuracy to the robustness of underlying reasoning.
Original Abstract
Class Activation Mapping (CAM) methods are widely used to generate visual explanations for deep learning classifiers in medical imaging. However, existing evaluation frameworks assess whether explanations are correct, measured by localisation fidelity against radiologist annotations, rather than whether they are consistent: whether the model applies the same spatial reasoning strategy across different patients with the same pathology. We propose the C-Score (Consistency Score), a confidence-weighted, annotation-free metric that quantifies intra-class explanation reproducibility via intensity-emphasised pairwise soft IoU across correctly classified instances. We evaluate six CAM techniques: GradCAM, GradCAM++, LayerCAM, EigenCAM, ScoreCAM, and MS GradCAM++ across three CNN architectures (DenseNet201, InceptionV3, ResNet50V2) over thirty training epochs on the Kermany chest X-ray dataset, covering transfer learning and fine-tuning phases. We identify three distinct mechanisms of AUC-consistency dissociation, invisible to standard classification metrics: threshold-mediated gold list collapse, technique-specific attribution collapse at peak AUC, and class-level consistency masking in global aggregation. C-Score provides an early warning signal of impending model instability. ScoreCAM deterioration on ResNet50V2 is detectable one full checkpoint before catastrophic AUC collapse and yields architecture-specific clinical deployment recommendations grounded in explanation quality rather than predictive ranking alone.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.