LARAG: Link-Aware Retrieval Strategy for RAG Systems in Hyperlinked Technical Documentation
Giorgia Bolognesi, Claudio Estatico, Ulderico Fugacci, Isabella Mastroianni, Claudio Muselli + 1 more
TLDR
LARAG improves RAG systems by leveraging hyperlink structures in technical documentation for more efficient and accurate content retrieval.
Key contributions
- LARAG is a link-aware retrieval strategy for RAG in hyperlinked technical documentation.
- It leverages existing HTML hyperlink topology by encoding relations as metadata in chunk representations.
- Enables implicit graph-like retrieval without explicit graph construction or inference.
- Consistently improves answer quality (BERTScore F1) while retrieving fewer chunks and tokens.
Why it matters
Current RAG systems often ignore the valuable hyperlink structure in technical documents, treating them as flat. LARAG addresses this by integrating link awareness directly into retrieval. This leads to more faithful and efficient RAG pipelines, improving answer quality and reducing computational costs for LLMs.
Original Abstract
Retrieval-Augmented Generation (RAG) enhances the factual grounding of Large Language Models by conditioning their outputs on external documents. However, standard embedding-based retrievers treat naturally structured corpora, such as technical manuals, as flat collections of passages, thereby overlooking the hyperlink topology that users rely on when navigating such content. We introduce LARAG (Link-Aware RAG): a lightweight, link-aware retrieval strategy that leverages the author-defined hyperlink structure already present in HTML documentation, encoding hyperlink relations as metadata in the chunk representations and exploiting them to perform a form of graph-like retrieval of locally relevant content. In a benchmark of twenty expert-designed queries over Rulex Platform technical documentation and four prompting strategies, LARAG consistently improves answer quality, achieving the highest BERTScore F1, while retrieving fewer chunks and generating fewer tokens than a baseline RAG architecture used for comparison. These results show that directly leveraging the existing hyperlink topology of technical documentation, even without explicit graph construction or inference, enables an implicit form of graph-like retrieval that yields a more faithful and efficient RAG pipeline, providing better grounding at lower cost.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.