ArXiv TLDR

Beyond Dense Connectivity: Explicit Sparsity for Scalable Recommendation

🐦 Tweet
2604.08011

Yantao Yu, Sen Qiao, Lei Shen, Bing Wang, Xiaoyi Zeng

cs.IR

TLDR

SSR introduces explicit sparsity to recommender systems, outperforming dense models by filtering low-utility connections for better scalability and performance.

Key contributions

  • Reveals "implicit connection sparsity" in dense recommenders, showing a structural mismatch with sparse input data.
  • Introduces SSR, a "filter-then-fuse" framework that explicitly integrates sparsity into the recommendation architecture.
  • Employs Static Random Filter and Iterative Competitive Sparse (ICS) for efficient, adaptive dimension filtering.
  • Outperforms state-of-the-art baselines and demonstrates superior scalability on large-scale industrial datasets.

Why it matters

Dense recommender models struggle with sparse data, leading to performance bottlenecks. This paper addresses this by explicitly integrating sparsity, a fundamental shift in architecture design. The proposed SSR framework not only achieves SOTA performance but also offers superior scalability, crucial for real-world industrial applications.

Original Abstract

Recent progress in scaling large models has motivated recommender systems to increase model depth and capacity to better leverage massive behavioral data. However, recommendation inputs are high-dimensional and extremely sparse, and simply scaling dense backbones (e.g., deep MLPs) often yields diminishing returns or even performance degradation. Our analysis of industrial CTR models reveals a phenomenon of implicit connection sparsity: most learned connection weights tend towards zero, while only a small fraction remain prominent. This indicates a structural mismatch between dense connectivity and sparse recommendation data; by compelling the model to process vast low-utility connections instead of valid signals, the dense architecture itself becomes the primary bottleneck to effective pattern modeling. We propose \textbf{SSR} (Explicit \textbf{S}parsity for \textbf{S}calable \textbf{R}ecommendation), a framework that incorporates sparsity explicitly into the architecture. SSR employs a multi-view "filter-then-fuse" mechanism, decomposing inputs into parallel views for dimension-level sparse filtering followed by dense fusion. Specifically, we realize the sparsity via two strategies: a Static Random Filter that achieves efficient structural sparsity via fixed dimension subsets, and Iterative Competitive Sparse (ICS), a differentiable dynamic mechanism that employs bio-inspired competition to adaptively retain high-response dimensions. Experiments on three public datasets and a billion-scale industrial dataset from AliExpress (a global e-commerce platform) show that SSR outperforms state-of-the-art baselines under similar budgets. Crucially, SSR exhibits superior scalability, delivering continuous performance gains where dense models saturate.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.