ShapShift: Explaining Model Prediction Shifts with Subgroup Conditional Shapley Values
Tom Bewley, Salim I. Amoukou, Emanuele Albini, Saumitra Mishra, Manuela Veloso
TLDR
ShapShift attributes machine learning model prediction shifts to changes in interpretable data subgroups using a novel Shapley value method.
Key contributions
- Proposes ShapShift, a Shapley value method to attribute prediction shifts to subgroup conditional probability changes.
- Provides exact explanations for single decision trees based on conditional probability changes at split nodes.
- Extends the method to tree ensembles, handling residual effects.
- Introduces a model-agnostic variant using surrogate trees, applicable to neural networks.
Why it matters
Prediction shifts impact business outcomes, making their causes crucial. ShapShift offers simple, faithful, and near-complete explanations across model classes. This improves model monitoring and helps diagnose issues in dynamic environments.
Original Abstract
Changes in input distribution can induce shifts in the average predictions of machine learning models. Such prediction shifts may impact downstream business outcomes (e.g. a bank's loan approval rate), so understanding their causes can be crucial. We propose \ours{}: a Shapley value method for attributing prediction shifts to changes in the conditional probabilities of interpretable subgroups of data, where these subgroups are defined by the structure of decision trees. We initially apply this method to single decision trees, providing exact explanations based on conditional probability changes at split nodes. Next, we extend it to tree ensembles by selecting the most explanatory tree and accounting for residual effects. Finally, we propose a model-agnostic variant using surrogate trees grown with a novel objective function, allowing application to models like neural networks. While exact computation can be intensive, approximation techniques enable practical application. We show that \ours{} provides simple, faithful, and near-complete explanations of prediction shifts across model classes, aiding model monitoring in dynamic environments.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.