ArXiv TLDR

One Pass, Any Order: Position-Invariant Listwise Reranking for LLM-Based Recommendation

🐦 Tweet
2604.27599

Ethan Bito, Yongli Ren, Estrid He

cs.IRcs.LG

TLDR

InvariRank is a new LLM-based reranking framework that ensures stable, order-independent recommendations in a single pass, addressing order sensitivity.

Key contributions

  • Addresses LLM reranker order sensitivity where candidate presentation affects rankings.
  • Proposes InvariRank, an architecture for permutation-invariant listwise reranking.
  • Achieves invariance via a structured attention mask and shared RoPE, blocking position effects.
  • Scores all candidates in one pass, avoiding costly permutation-based invariance training.

Why it matters

LLM-based rerankers are unreliable due to their dependence on candidate order. InvariRank offers a practical, efficient, and architecturally sound solution, making LLM recommendations more stable and trustworthy. This enhances the reliability and efficiency of LLM-based recommendation systems.

Original Abstract

Large language models (LLMs) are increasingly used for recommendation reranking, but their listwise predictions can depend on the order in which candidates are presented. This creates a mismatch between the set-based nature of recommendation and the sequence-based computation of decoder-only LLMs, where permuting an otherwise identical candidate set can change item scores and final rankings. Such order sensitivity makes LLM-based rerankers difficult to rely on, since rankings may reflect prompt serialization rather than user preference. We propose InvariRank, a permutation-invariant listwise reranking framework that addresses this dependence at the architectural level. InvariRank blocks cross-candidate attention with a structured attention mask and negates position-induced scoring changes through shared positional framing under Rotary Positional Embeddings (RoPE). Combined with a listwise learning-to-rank objective, the model scores all candidates in a single forward pass, avoiding permutation-based invariance training objectives that require multiple permutations of a candidate set. Experiments on recommendation benchmarks show that InvariRank maintains competitive ranking effectiveness while producing stable rankings across candidate permutations. The results suggest that architectural invariance is a practical route to reliable and efficient LLM-based recommendation reranking. The source code is at https://github.com/ejbito/InvariRank.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.