Multi-LLM Token Filtering and Routing for Sequential Recommendation
Wuhan Chen, Min Gao, Xin Xia, Zongwei Wang, Wentao Li + 1 more
TLDR
MLTFR improves sequential recommendation by filtering and routing token embeddings from multiple LLMs without external textual input.
Key contributions
- Proposes MLTFR, a corpus-free framework for sequential recommendation using only LLM token embeddings.
- Introduces user-guided token filtering to select task-relevant embeddings and suppress noisy vocabulary signals.
- Integrates multiple LLM token spaces via a Mixture-of-Experts architecture for comprehensive semantic knowledge.
- Employs a Fisher-weighted semantic consensus expert to balance diverse LLM contributions effectively.
Why it matters
This paper addresses a fundamental challenge in leveraging LLMs for recommendation without relying on external text. By proposing MLTFR, it offers a novel, corpus-free approach that effectively integrates and filters knowledge from multiple LLMs. This leads to more stable and superior sequential recommendation performance, opening new avenues for LLM application in this domain.
Original Abstract
Large language models (LLMs) have recently shown promise in recommendation by providing rich semantic knowledge. While most existing approaches rely on external textual corpora to align LLMs with recommender systems, we revisit a more fundamental yet underexplored question: Can recommendation benefit from LLM token embeddings alone without textual input? Through a systematic empirical study, we show that directly injecting token embeddings from a single LLM into sequential recommenders leads to unstable or limited gains, due to semantic misalignment, insufficient task adaptation, and the restricted coverage of individual LLMs. To address these challenges, we propose MLTFR, a Multi-LLM Token Filtering and Routing framework for corpus-free sequential recommendation. MLTFR follows an interaction-guided LLM knowledge integration paradigm, where task-relevant token embeddings are selected via user-guided token filtering to suppress noisy and irrelevant vocabulary signals. To overcome the limitations of single-LLM representations, MLTFR integrates multiple LLM token spaces through a Mixture-of-Experts architecture, with a Fisher-weighted semantic consensus expert to balance heterogeneous experts and prevent domination during training. By jointly filtering informative tokens and aggregating complementary semantic knowledge across multiple LLMs, MLTFR enables stable and effective utilization of LLM token embeddings without textual inputs or backbone modification. Extensive experiments demonstrate that MLTFR consistently outperforms state-of-the-art sequential recommendation baselines and existing alignment methods. Our code is available at: https://github.com/ccwwhhh/MLTFR.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.