ArXiv TLDR

Metric-agnostic Learning-to-Rank via Boosting and Rank Approximation

🐦 Tweet
2604.15101

Camilo Gomez, Pengyang Wang, Yanjie Fu

cs.IRcs.LG

TLDR

This paper introduces a novel metric-agnostic Learning-to-Rank framework that uses a differentiable loss and gradient boosting for improved, generalizable ranking.

Key contributions

  • Proposes a novel listwise Learning-to-Rank (LTR) framework for efficient and generalizable ranking.
  • Introduces a new differentiable ranking loss combining smooth rank approximation with average mean square loss.
  • Adapts gradient-boosting machines to efficiently minimize the proposed listwise ranking loss.
  • Achieves state-of-the-art performance across various IR metrics with comparable computational efficiency.

Why it matters

Current LTR methods struggle with non-differentiable optimization and limited generalization across different ranking metrics. This paper addresses these issues by proposing a novel, metric-agnostic framework. It offers a more stable and efficient training process, leading to better performance and broader applicability in information retrieval systems.

Original Abstract

Learning-to-Rank (LTR) is a supervised machine learning approach that constructs models specifically designed to order a set of items or documents based on their relevance or importance to a given query or context. Despite significant success in real-world information retrieval systems, current LTR methods rely on one prefix ranking metric (e.g., such as Normalized Discounted Cumulative Gain (NDCG) or Mean Average Precision (MAP)) for optimizing the ranking objective function. Such metric-dependent setting limits LTR methods from two perspectives: (1) non-differentiable problem: directly optimizing ranking functions over a given ranking metric is inherently non-smooth, making the training process unstable and inefficient; (2) limited ranking utility: optimizing over one single metric makes it difficult to generalize well to other ranking metrics of interest. To address the above issues, we propose a novel listwise LTR framework for efficient and generalizable ranking purpose. Specifically, we propose a new differentiable ranking loss that combines a smooth approximation to the ranking operator with the average mean square loss per query. Then, we adapt gradient-boosting machines to minimize our proposed loss with respect to each list, a novel contribution. Finally, extensive experimental results confirm that our method outperforms the current state-of-the-art in information retrieval measures with similar efficiency.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.