Figures as Interfaces: Toward LLM-Native Artifacts for Scientific Discovery
Yifang Wang, Rui Sheng, Erzhuo Shao, Yifan Qian, Haotian Li + 2 more
TLDR
This paper introduces LLM-native figures, data-driven artifacts that enable LLMs to interact with and extend scientific visualizations by embedding full provenance.
Key contributions
- Introduces "LLM-native figures" that are both human-legible and machine-addressable for scientific discovery.
- Embeds complete provenance (data, code, viz spec) directly into figures, allowing LLMs to "see through" them.
- Enables LLMs to trace selections, generate code, and orchestrate new visualizations via natural language.
- Implements a hybrid language-visual interface, demonstrating accelerated discovery and improved reproducibility.
Why it matters
This paper redefines scientific figures from static outputs to interactive interfaces for LLMs. By embedding full provenance, it allows LLMs to deeply engage with data and analyses, fostering more dynamic and reproducible scientific workflows. This approach promises to accelerate discovery and enhance transparency in human-AI collaboration.
Original Abstract
Large language models (LLMs) are transforming scientific workflows, not only through their generative capabilities but also through their emerging ability to use tools, reason about data, and coordinate complex analytical tasks. Yet in most human-AI collaborations, the primary outputs, figures, are still treated as static visual summaries: once rendered, they are handled by both humans and multimodal LLMs as images to be re-interpreted from pixels or captions. The emergent capabilities of LLMs open an opportunity to fundamentally rethink this paradigm. In this paper, we introduce the concept of LLM-native figures: data-driven artifacts that are simultaneously human-legible and machine-addressable. Unlike traditional plots, each artifact embeds complete provenance: the data subset, analytical operations and code, and visualization specification used to generate it. As a result, an LLM can "see through" the figure--tracing selections back to their sources, generating code to extend analyses, and orchestrating new visualizations through natural-language instructions or direct manipulation. We implement this concept through a hybrid language-visual interface that integrates LLM agents with a bidirectional mapping between figures and underlying data. Using the science of science domain as a testbed, we demonstrate that LLM-native figures can accelerate discovery, improve reproducibility, and make reasoning transparent across agents and users. More broadly, this work establishes a general framework for embedding provenance, interactivity, and explainability into the artifacts of modern research, redefining the figure not as an end product, but as an interface for discovery. For more details, please refer to the demo video available at www.llm-native-figure.com.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.