ArXiv TLDR

PREF-XAI: Preference-Based Personalized Rule Explanations of Black-Box Machine Learning Models

🐦 Tweet
2604.19684

Salvatore Greco, Jacek Karolczak, Roman Słowiński, Jerzy Stefanowski

cs.LG

TLDR

PREF-XAI introduces a novel preference-based framework for personalized rule explanations of black-box models, learning user preferences from limited feedback.

Key contributions

  • Proposes PREF-XAI, reframing explanation as a preference-driven decision problem.
  • Combines rule-based explanations with formal preference learning using robust ordinal regression.
  • Elicits user preferences via ranking and models them with an additive utility function.
  • Accurately reconstructs preferences, identifies relevant explanations, and discovers novel rules.

Why it matters

This paper addresses a key limitation in XAI by personalizing explanations based on individual user preferences. It offers a principled approach to learn and apply these preferences, leading to more relevant and adaptive explanation systems. This connection between XAI and preference learning opens new research avenues.

Original Abstract

Explainable artificial intelligence (XAI) has predominantly focused on generating model-centric explanations that approximate the behavior of black-box models. However, such explanations often overlook a fundamental aspect of interpretability: different users require different explanations depending on their goals, preferences, and cognitive constraints. Although recent work has explored user-centric and personalized explanations, most existing approaches rely on heuristic adaptations or implicit user modeling, lacking a principled framework for representing and learning individual preferences. In this paper, we consider Preference-Based Explainable Artificial Intelligence (PREF-XAI), a novel perspective that reframes explanation as a preference-driven decision problem. Within PREF-XAI, explanations are not treated as fixed outputs, but as alternatives to be evaluated and selected according to user-specific criteria. In the PREF-XAI perspective, here we propose a methodology that combines rule-based explanations with formal preference learning. User preferences are elicited through a ranking of a small set of candidate explanations and modeled via an additive utility function inferred using robust ordinal regression. Experimental results on real-world datasets show that PREF-XAI can accurately reconstruct user preferences from limited feedback, identify highly relevant explanations, and discover novel explanatory rules not initially considered by the user. Beyond the proposed methodology, this work establishes a connection between XAI and preference learning, opening new directions for interactive and adaptive explanation systems.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.