XGRAG: A Graph-Native Framework for Explaining KG-based Retrieval-Augmented Generation
Zhuoling Li, Ha Linh Hong Tran Nguyen, Valeria Bladinieres, Maxim Romanovsky
TLDR
XGRAG introduces a novel framework for explaining GraphRAG systems, using graph perturbations to quantify knowledge graph component contributions.
Key contributions
- Introduces XGRAG, a framework for causally grounded explanations in GraphRAG systems.
- Uses graph-based perturbation strategies to quantify individual graph component contributions.
- Achieves 14.81% better explanation quality than RAG-Ex baseline across multiple QA datasets.
- Explanations correlate with graph centrality, validating its ability to capture graph structure.
Why it matters
GraphRAG systems offer grounded LLM answers but lack transparency regarding their reasoning. XGRAG addresses this by providing interpretable, causally grounded explanations for how knowledge graph components influence outputs. This enhances trustworthiness and interpretability of advanced RAG systems, fostering more reliable AI.
Original Abstract
Graph-based Retrieval-Augmented Generation (GraphRAG) extends traditional RAG by using knowledge graphs (KGs) to give large language models (LLMs) a structured, semantically coherent context, yielding more grounded answers. However, GraphRAG reasoning process remains a black-box, limiting our ability to understand how specific pieces of structured knowledge influence the final output. Existing explainability (XAI) methods for RAG systems, designed for text-based retrieval, are limited to interpreting an LLM response through the relational structures among knowledge components, creating a critical gap in transparency and trustworthiness. To address this, we introduce XGRAG, a novel framework that generates causally grounded explanations for GraphRAG systems by employing graph-based perturbation strategies, to quantify the contribution of individual graph components on the model answer. We conduct extensive experiments comparing XGRAG against RAG-Ex, an XAI baseline for standard RAG, and evaluate its robustness across various question types, narrative structures and LLMs. Our results demonstrate a 14.81% improvement in explanation quality over the baseline RAG-Ex across NarrativeQA, FairyTaleQA, and TriviaQA, evaluated by F1-score measuring alignment between generated explanations and original answers. Furthermore, XGRAG explanations exhibit a strong correlation with graph centrality measures, validating its ability to capture graph structure. XGRAG provides a scalable and generalizable approach towards trustworthy AI through transparent, graph-based explanations that enhance the interpretability of RAG systems.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.