UniRec: Bridging the Expressive Gap between Generative and Discriminative Recommendation via Chain-of-Attribute
Ziliang Wang, Gaoyun Lin, Xuesi Wang, Shaoqiang Liang, Yili Huang + 1 more
TLDR
UniRec bridges the expressive gap in generative recommendation by using Chain-of-Attribute to incorporate item features, outperforming baselines significantly.
Key contributions
- UniRec uses Chain-of-Attribute (CoA) to integrate item features (category, seller, brand) into generative recommendation.
- CoA recovers item-side feature crossing, reducing decoding entropy and stabilizing beam search trajectories.
- Introduces Capacity-constrained SID and Conditional Decoding Context (CDC) for robust deployment.
- Outperforms baselines by +22.6% HR@50 and shows significant online business metric gains.
Why it matters
Generative recommendation offers a unified pipeline but has an expressive gap due to limited item-side feature access. UniRec addresses this by integrating item attributes via Chain-of-Attribute, making generative models as expressive as discriminative ones. This significantly improves performance and practical deployment, advancing the field.
Original Abstract
Generative Recommendation (GR) reframes retrieval and ranking as autoregressive decoding over Semantic IDs (SIDs), unifying the multi-stage pipeline into a single model. Yet a fundamental expressive gap persists: discriminative models score items with direct feature access, enabling explicit user-item crossing, whereas GR decodes over compact SID tokens without item-side signal. We formalize this via Bayes' theorem, showing ranking by p(y|f,u) is equivalent to ranking by p(f|y,u), which factorizes autoregressively over item features. This establishes that a generative model with full feature access is as expressive as its discriminative counterpart; any practical gap stems solely from incomplete feature coverage. We propose UniRec with Chain-of-Attribute (CoA) as its core mechanism. CoA prefixes each SID sequence with structured attribute tokens--category, seller, brand--before decoding the SID itself, recovering the item-side feature crossing that discriminative models exploit. Because items sharing identical attributes cluster in adjacent SID regions, attribute conditioning yields a measurable per-step entropy reduction H(s_k|s_{<k},a) < H(s_k|s_{<k}), narrowing the search space and stabilizing beam search trajectories. We further address two deployment challenges: Capacity-constrained SID introduces exposure-weighted capacity penalties into residual quantization to suppress token collapse and the Matthew effect across SID layers; Conditional Decoding Context (CDC) combines Task-Conditioned BOS with hash-based Content Summaries, injecting scenario-conditioned signals at each decoding step. A joint RFT and DPO framework aligns the model with business objectives beyond distribution matching. Experiments show UniRec outperforms the strongest baseline by +22.6% HR@50 overall and +15.5% on high-value orders, with online A/B tests confirming significant business metric gains.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.