ArXiv TLDR

Explainable AI for Jet Tagging: A Comparative Study of GNNExplainer, GNNShap, and GradCAM for Jet Tagging in the Lund Jet Plane

🐦 Tweet
2604.25885

Pahal D. Patel, Sanmay Ganguly

hep-phcs.LGhep-ex

TLDR

This paper compares GNNExplainer, GNNShap, and GradCAM for explainable AI in jet tagging using LundNet's graph representation.

Key contributions

  • Adapted GNNExplainer, GNNShap, and GradCAM for explainable jet tagging on LundNet's graph.
  • Developed a physics-informed evaluation framework using Monte Carlo truth explanation masks.
  • Analyzed explanation quality and focus across different transverse-momentum bins.
  • Quantified correlation between explainer importance and classical jet substructure observables.

Why it matters

This paper addresses the opacity of state-of-the-art GNNs in jet tagging by providing methods to explain their predictions. It links model explanations to physical reasoning and known QCD features, enhancing trust and guiding the development of more interpretable and physically-informed models.

Original Abstract

Graph neural networks such as ParticleNet and transformer based networks on point clouds such as ParticleTransformer achieve state-of-the-art performance on jet tagging benchmarks at the Large Hadron Collider, yet the physical reasoning behind their predictions remains opaque. We present different methods, i.e. perturbation-based (GNNExplainer), Shapley-value-based (GNNShap), and gradient-based (GRADCam); adapted to operate on LundNet's Lund-plane graph representation. Leveraging the fact that each node in the Lund plane corresponds to a physically meaningful parton splitting, we construct Monte Carlo truth explanation masks and introduce a physics-informed evaluation framework that goes beyond standard fidelity metrics. We perform the analysis in three transverse-momentum bins ($\mathrm{p_T} \in [500,700]$, $[800,1000]$, and the inclusive region $[500,1000]$ GeV), revealing how explanation quality and focus shift between non-perturbative and perturbative regimes. We further quantify the correlation between explainer-assigned node importance and classical jet substructure observables -- $N$-subjettiness ratios $τ_{21}$ and $τ_{32}$ and the energy correlation functions -- establishing the degree to which the model has learned known QCD features. We find that overall the weight assigned by explainability methods has a correlation with analytic observables, with expected shift across different phase space regimes, indicating that a trained neural network indeed learns some aspects of jet-substructure moments. Our open-source implementation enables reproducible explainability studies for graph-based jet taggers.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.