ArXiv TLDR

ShapShift: Explaining Model Prediction Shifts with Subgroup Conditional Shapley Values

🐦 Tweet
2604.11200

Tom Bewley, Salim I. Amoukou, Emanuele Albini, Saumitra Mishra, Manuela Veloso

cs.LGcs.AIstat.ML

TLDR

ShapShift attributes machine learning model prediction shifts to changes in interpretable data subgroups using a novel Shapley value method.

Key contributions

  • Proposes ShapShift, a Shapley value method to attribute prediction shifts to subgroup conditional probability changes.
  • Provides exact explanations for single decision trees based on conditional probability changes at split nodes.
  • Extends the method to tree ensembles, handling residual effects.
  • Introduces a model-agnostic variant using surrogate trees, applicable to neural networks.

Why it matters

Prediction shifts impact business outcomes, making their causes crucial. ShapShift offers simple, faithful, and near-complete explanations across model classes. This improves model monitoring and helps diagnose issues in dynamic environments.

Original Abstract

Changes in input distribution can induce shifts in the average predictions of machine learning models. Such prediction shifts may impact downstream business outcomes (e.g. a bank's loan approval rate), so understanding their causes can be crucial. We propose \ours{}: a Shapley value method for attributing prediction shifts to changes in the conditional probabilities of interpretable subgroups of data, where these subgroups are defined by the structure of decision trees. We initially apply this method to single decision trees, providing exact explanations based on conditional probability changes at split nodes. Next, we extend it to tree ensembles by selecting the most explanatory tree and accounting for residual effects. Finally, we propose a model-agnostic variant using surrogate trees grown with a novel objective function, allowing application to models like neural networks. While exact computation can be intensive, approximation techniques enable practical application. We show that \ours{} provides simple, faithful, and near-complete explanations of prediction shifts across model classes, aiding model monitoring in dynamic environments.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.