ArXiv TLDR

VISTA: Visualization of Token Attribution via Efficient Analysis

🐦 Tweet
2604.02217

Syed Ahmed, Bharathi Vokkaliga Ganesh, Jagadish Babu P, Karthick Selvaraj, Praneeth Talluri + 6 more

cs.AIcs.CL

TLDR

VISTA introduces a model-agnostic, perturbation-based method to visualize token importance in LLMs without extra computational cost, using a three-matrix framework.

Key contributions

  • Proposes VISTA, a model-agnostic technique for visualizing token importance in LLMs.
  • Leverages perturbation-based strategies without backpropagation, reducing GPU memory and cost.
  • Introduces a novel three-matrix analytical framework: Angular, Magnitude, and Dimensional Deviation Matrices.
  • Derives a composite importance score for nuanced and mathematically grounded token significance.

Why it matters

This paper matters by offering a lightweight, model-agnostic solution to interpret LLM behavior, overcoming limitations of prior architecture-specific and computationally expensive methods. It enhances transparency and understanding of generative AI systems, making their internal workings more accessible.

Original Abstract

Understanding how Large Language Models (LLMs) process information from prompts remains a significant challenge. To shed light on this "black box," attention visualization techniques have been developed to capture neuron-level perceptions and interpret how models focus on different parts of input data. However, many existing techniques are tailored to specific model architectures, particularly within the Transformer family, and often require backpropagation, resulting in nearly double the GPU memory usage and increased computational cost. A lightweight, model-agnostic approach for attention visualization remains lacking. In this paper, we introduce a model-agnostic token importance visualization technique to better understand how generative AI systems perceive and prioritize information from input text, without incurring additional computational cost. Our method leverages perturbation-based strategies combined with a three-matrix analytical framework to generate relevance maps that illustrate token-level contributions to model predictions. The framework comprises: (1) the Angular Deviation Matrix, which captures shifts in semantic direction; (2) the Magnitude Deviation Matrix, which measures changes in semantic intensity; and (3) the Dimensional Importance Matrix, which evaluates contributions across individual vector dimensions. By systematically removing each token and measuring the resulting impact across these three complementary dimensions, we derive a composite importance score that provides a nuanced and mathematically grounded measure of token significance. To support reproducibility and foster wider adoption, we provide open-source implementations of all proposed and utilized explainability techniques, with code and resources publicly available at https://github.com/Infosys/Infosys-Responsible-AI-Toolkit

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.