UniRank: Unified List-wise Reranking via Confidence-Ordered Denoising
Pengyue Jia, Hailan Yang, Shuchang Liu, Xiaobei Wang, Wanyu Wang + 5 more
TLDR
UniRank unifies autoregressive and non-autoregressive reranking using confidence-ordered denoising, improving performance and user engagement.
Key contributions
- Unifies autoregressive (AR) and non-autoregressive (NAR) reranking, combining their strengths.
- Uses iterative denoising with bidirectional slate modeling to fill slots based on confidence.
- Introduces Task Grounded Diffusion Interface (TGD) for item-level denoising in candidate pools.
- Outperforms SOTA baselines and shows significant gains in online A/B tests.
Why it matters
Existing rerankers suffer from error propagation or weak inter-item modeling. UniRank offers a unified solution, combining the best of both worlds with a novel denoising approach. This leads to superior performance and tangible improvements in user engagement on real-world platforms.
Original Abstract
List-wise reranking arranges a request-specific pool of candidate items into an ordered slate that maximizes user satisfaction. Existing generative rerankers fall into two paradigms: Autoregressive (AR) rerankers construct the slate left to right and capture inter-item dependencies in the exposure list, but they suffer from error propagation because early mistakes affect subsequent slots. Non-autoregressive (NAR) rerankers predict all slots in parallel and avoid error propagation, but they weaken inter-item interaction modeling under a slot independence assumption. This raises a central question: is there a unified architecture that combines the strengths of both paradigms and delivers stronger reranking performance? We answer this question with UniRank, a unified list-wise reranking framework whose inference time variants recover AR and NAR rerankers as special cases. UniRank integrates bidirectional slate modeling into an iterative denoising process and fills the most confident slot at each step. To instantiate this framework for reranking, we introduce the Task Grounded Diffusion Interface (TGD), which performs denoising at the item level and restricts prediction to the request-specific candidate pool. TGD aggregates each item's semantic tokens into a single item embedding and scores each slot directly against the candidate pool. Experiments on Amazon Books, MovieLens-1M, and an industrial short video dataset show that UniRank consistently outperforms state-of-the-art baselines. Online A/B tests on a real-world industrial platform further validate its effectiveness, yielding significant improvements of +0.159% in user average app-time and +1.016% in share-rate.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.