ArXiv TLDR

Factorized Latent Reasoning for LLM-based Recommendation

🐦 Tweet
2604.26760

Tianqi Gao, Chengkai Huang, Zihan Wang, Cao Liu, Ke Zeng + 1 more

cs.IR

TLDR

This paper introduces Factorized Latent Reasoning (FLR), an LLM-based recommendation framework that disentangles user preferences into multiple factors.

Key contributions

  • Decomposes user preferences into multiple disentangled latent factors for better representation.
  • Employs a lightweight multi-factor attention module for iterative refinement of latent thoughts.
  • Utilizes orthogonality, diversity, and sparsity regularization for factor specialization.
  • Integrates reinforcement learning for stable alignment in the latent reasoning space.

Why it matters

Existing LLM recommenders struggle with complex user preferences due to single latent vectors. FLR addresses this by providing a more nuanced, multi-faceted approach. This improves recommendation accuracy, robustness, and interpretability, making LLM-based systems more effective.

Original Abstract

Large language models (LLMs) have recently been adopted for recommendation by framing user preference modeling as a language generation problem. However, existing latent reasoning approaches typically represent user intent with a single latent vector, which struggles to capture the inherently multi-faceted nature of user preferences. We propose Factorized Latent Reasoning (FLR), a novel framework for LLM-based sequential recommendation that decomposes latent reasoning into multiple disentangled preference factors. FLR introduces a lightweight multi-factor attention module that iteratively refines a latent thought representation, where each factor attends to distinct aspects of the user's interaction history. To encourage diversity and specialization, we design orthogonality, attention diversity, and sparsity regularization objectives, and dynamically aggregate factor contributions for the final prediction. We further integrate FLR with an efficient reinforcement learning strategy based on group-relative policy optimization, enabling stable alignment directly in the latent reasoning space. Experiments on multiple benchmarks show that FLR consistently outperforms strong baselines while improving robustness and interpretability.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.