R3-VAE: Reference Vector-Guided Rating Residual Quantization VAE for Generative Recommendation
Qiang Wan, Ze Yang, Dawei Yang, Ying Fan, Xin Yan + 1 more
TLDR
R3-VAE enhances generative recommendation by using reference vectors and novel rating mechanisms to stabilize training and improve semantic identifier quality.
Key contributions
- Introduces a reference vector as a semantic anchor to mitigate initialization sensitivity in SID generation.
- Employs a dot product-based rating mechanism for stable training and to prevent codebook collapse.
- Proposes two new SID evaluation metrics, Semantic Cohesion and Preference Discrimination, for regularization.
- Achieves significant performance gains (e.g., 14.2% Recall@10) and improves cold-start in real-world systems.
Why it matters
This paper tackles key challenges in generative recommendation, improving the stability of semantic identifier training and providing better quality assessment. Its innovations lead to more robust and effective recommendation systems, with significant performance gains and real-world industrial applicability, especially for cold-start scenarios.
Original Abstract
Generative Recommendation (GR) has gained traction for its merits of superior performance and cold-start capability. As the vital role in GR, Semantic Identifiers (SIDs) represent item semantics through discrete tokens. However, current techniques for SID generation based on vector quantization face two main challenges: (i) training instability, stemming from insufficient gradient propagation through the straight-through estimator and sensitivity to initialization; and (ii) inefficient SID quality assessment, where industrial practice still depends on costly GR training and A/B testing. To address these challenges, we propose Reference Vector-Guided Rating Residual Quantization VAE (R3-VAE). This framework incorporates three key innovations: (i) a reference vector that functions as a semantic anchor for the initial features, thereby mitigating sensitivity to initialization; (ii) a dot product-based rating mechanism designed to stabilize the training process and prevent codebook collapse; and (iii) two SID evaluation metrics, Semantic Cohesion and Preference Discrimination, serving as regularization terms during training. Empirical results on six benchmarks demonstrate that R3-VAE outperforms state-of-the-art methods, achieving an average improvement of 14.2% in Recall@10 and 15.5% in NDCG@10 across three Amazon datasets. Furthermore, we perform GR training and online A/B tests on a prominent news recommendation platform. Our method achieves a 1.62% improvement in MRR and a 0.83% gain in StayTime/U versus baselines. Additionally, we employ R3-VAE to replace the item ID of CTR model, resulting in significant improvements in content cold start by 15.36%, corroborating the strong applicability and business value in industry-scale recommendation scenarios.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.