ArXiv TLDR

LLM-Augmented Semantic Steering of Text Embedding Projection Spaces

🐦 Tweet
2605.01957

Wei Liu, Eric Krokos, Kirsten Whitley, Rebecca Faust, Chris North

cs.HCcs.CL

TLDR

This paper introduces LLM-augmented semantic steering to reorganize text embedding projections based on user intent, improving alignment without retraining.

Key contributions

  • Users express semantic intent by grouping example documents within projections.
  • LLMs externalize this intent as natural language and extend it to related documents.
  • Semantic info is incorporated via text augmentation or embedding blending, without retraining.
  • Improves alignment with target semantic structures using minimal user interaction.

Why it matters

Existing methods for text embedding projections lack interpretability and flexibility. This paper introduces an explicit, language-mediated interaction method, transforming static projections into dynamic, intent-dependent semantic workspaces. This significantly enhances the visual analysis of document collections.

Original Abstract

Low-dimensional projections of text embeddings support visual analysis of document collections, but their spatial organization may not reflect the relationships an analyst intends to examine. Existing semantic interaction approaches encode semantic intent indirectly through geometric constraints or model updates, limiting interpretability and flexibility. We introduce LLM-augmented semantic steering, which enables analysts to express semantic intent by grouping a small set of example documents within the projection. A large language model externalizes this intent as natural-language representations and selectively extends it to related documents; the resulting semantic information is then incorporated into document representations via text augmentation or embedding-level blending, without retraining the underlying models. A case study illustrates how the same corpus can be reorganized from different semantic perspectives, while simulation-based evaluation shows that semantic steering improves global and local alignment with target semantic structures using only minimal interaction. Embedding-level blending further enables continuous and controllable steering of projection layouts. These results position projection spaces as intent-dependent semantic workspaces that can be reshaped through explicit, interpretable, language-mediated interaction.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.