ArXiv TLDR

RRCM: Ranking-Driven Retrieval over Collaborative and Meta Memories for LLM Recommendation

🐦 Tweet
2605.07129

Shijun Li, Wooseong Yang, Yu Wang, Tianxin Wei, Joydeep Ghosh

cs.IRcs.AIcs.LG

TLDR

RRCM is a ranking-driven retrieval framework for LLM recommenders that dynamically selects collaborative and metadata evidence to improve recommendation quality.

Key contributions

  • Introduces RRCM, a framework for LLM recommenders to dynamically retrieve relevant collaborative and metadata evidence.
  • Overcomes fixed context construction and context window limitations by learning what information to retrieve.
  • Represents both collaborative and metadata memories in natural language, accessed via a unified interface.
  • Optimizes the memory-reading policy using an outcome-only ranking reward for improved top-k recommendations.

Why it matters

LLM recommenders struggle with fixed context construction and context window limits when using diverse data. RRCM introduces a novel, ranking-driven approach to dynamically retrieve and integrate collaborative and metadata evidence. This significantly improves recommendation accuracy and efficiency by learning what information is truly beneficial.

Original Abstract

Large Language Models (LLMs) have emerged as a promising paradigm for next-generation recommender systems, offering strong semantic understanding and natural-language reasoning abilities. Despite recent progress, current LLM-based recommenders still face key challenges in constructing decision-relevant contexts from heterogeneous evidence. First, existing methods often rely on fixed context construction strategies: collaborative behavioral evidence and item-side metadata are typically incorporated through predefined prompts, static retrieval pipelines, or handcrafted injection mechanisms, making it difficult to determine what information is truly beneficial for each instance. Second, heterogeneous evidence introduces a severe context-efficiency bottleneck. Rich metadata and collaborative interaction records can quickly overwhelm the context window, while aggressive compression or heuristic filtering may discard fine-grained evidence critical for accurate recommendation. To address these challenges, we propose RRCM, a ranking-driven retrieval-and-reasoning framework over collaborative and metadata memories for LLM-based agentic recommendation. RRCM starts from a lightweight user-history context and learns whether to recommend directly, retrieve collaborative evidence, retrieve item metadata, or interleave both through reasoning. Both memories are represented in natural language and accessed through a unified retrieval interface, enabling flexible evidence acquisition without handcrafted CF injection or fixed retrieval rules. We optimize this memory-reading policy with an outcome-only ranking reward, instantiated using group relative policy optimization, so that retrieval decisions are directly driven by final top-k recommendation quality. Extensive experiments show that RRCM significantly outperforms traditional baselines and diverse LLM-based recommendation approaches.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.