ArXiv TLDR

Superintelligent Retrieval Agent: The Next Frontier of Information Retrieval

🐦 Tweet
2605.06647

Zeyu Yang, Qi Ma, Jason Chen, Anshumali Shrivastava

cs.IRcs.AIcs.LG

TLDR

SIRA introduces a superintelligent retrieval agent that uses LLM-guided lexical queries and corpus statistics to achieve superior, efficient, single-round information retrieval.

Key contributions

  • Compresses multi-round exploratory search into a single, corpus-discriminative retrieval action.
  • Uses an LLM to enrich documents offline and predict query evidence vocabulary for better term selection.
  • Leverages document-frequency statistics to filter proposed terms, ensuring relevance and discriminative power.
  • Achieves superior performance over dense retrievers and multi-round agents with a single weighted BM25 call.

Why it matters

This paper introduces a novel, efficient approach to information retrieval that significantly improves upon existing multi-round agentic methods. By integrating LLM cognition with lightweight corpus statistics, SIRA offers a training-free, interpretable, and highly effective solution. This advancement could revolutionize how agents interact with large knowledge bases, reducing latency and improving recall.

Original Abstract

Retrieval-augmented agents are increasingly the interface to large organizational knowledge bases, yet most still treat retrieval as a black box: they issue exploratory queries, inspect returned snippets, and iteratively reformulate until useful evidence emerges. This approach resembles how a newcomer searches an unfamiliar database rather than how an expert navigates it with strong priors about terminology and likely evidence, and results in unnecessary retrieval rounds, increased latency, and poor recall. We introduce \textit{SuperIntelligent Retrieval Agent} (SIRA), which defines \emph{superintelligence} in retrieval as the ability to compress multi-round exploratory search into a single corpus-discriminative retrieval action. SIRA does not merely ask what terms are relevant to the query; it asks which terms are likely to separate the desired evidence from corpus-level confusers. On the corpus side, an LLM enriches each document offline with missing search vocabulary; on the query side, it predicts evidence vocabulary omitted by the query; and document-frequency statistics as a tool call to filter proposed terms that are absent, overly common, or unlikely to create retrieval margin. The final retrieval step is a single weighted BM25 call combining the original query with the validated expansion. Across ten BEIR benchmarks and downstream question-answering tasks, SIRA achieves the significantly superior performance outperforming dense retrievers and state-of-the-art multi-round agentic baselines, demonstrating that one well-formed lexical query, guided by LLM cognition and lightweight corpus statistics, can exceed substantially more expensive multi-round search while remaining interpretable, training-free, and efficient.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.