Dynamic Ranked List Truncation for Reranking Pipelines via LLM-generated Reference-Documents
Nilanjan Sinhababu, Soumedhik Bharati, Debasis Ganguly, Pabitra Mitra
TLDR
LLM-generated reference documents enable dynamic ranked list truncation and efficient listwise reranking, significantly accelerating LLM-based search pipelines.
Key contributions
- Uses LLM-generated "reference documents" to dynamically truncate ranked lists, acting as a relevance pivot.
- Enhances efficient listwise reranking frameworks by integrating these LLM-generated reference documents.
- Implements parallel batch processing with adaptive strides for faster, more efficient LLM-based reranking.
Why it matters
This paper significantly improves the efficiency of LLM-based reranking by tackling computational overhead and context length challenges. It offers a practical new paradigm for faster search, demonstrating LLMs' ability to generate semantically controlled documents.
Original Abstract
Large Language Models (LLM) have been widely used in reranking. Computational overhead and large context lengths remain a challenging issue for LLM rerankers. Efficient reranking usually involves selecting a subset of the ranked list from the first stage, known as ranked list truncation (RLT). The truncated list is processed further by a reranker. For LLM rerankers, the ranked list is often partitioned and processed sequentially in batches to reduce the context length. Both these steps involve hyperparameters and topic-agnostic heuristics. Recently, LLMs have been shown to be effective for relevance judgment. Equivalently, we propose that LLMs can be used to generate reference documents that can act as a pivot between relevant and non-relevant documents in a ranked list. We propose methods to use these generated reference documents for RLT as well as for efficient listwise reranking. While reranking, we process the ranked list in either parallel batches of non-overlapping windows or overlapping windows with adaptive strides, improving the existing fixed stride setup. The generated reference documents are also shown to improve existing efficient listwise reranking frameworks. Experiments on TREC Deep Learning benchmarks show that our approach outperforms existing RLT-based approaches. In-domain and out-of-domain benchmarks demonstrate that our proposed methods accelerate LLM-based listwise reranking by up to 66\% compared to existing approaches. This work not only establishes a practical paradigm for efficient LLM-based reranking but also provides insight into the capability of LLMs to generate semantically controlled documents using relevance signals.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.