ArXiv TLDR

Sparse Contrastive Learning for Content-Based Cold Item Recommendation

🐦 Tweet
2604.12990

Gregor Meehan, Johan Pauwels

cs.IR

TLDR

This paper introduces SEMCo, a novel content-based cold-start recommendation system using sparse contrastive learning with α-entmax to improve accuracy and equity.

Key contributions

  • Proposes a purely content-based model for cold-start recommendations, avoiding CF embedding alignment.
  • Introduces Sampled Entmax for Cold-start (SEMCo), using sparse contrastive learning with α-entmax.
  • SEMCo sharpens relevance estimation by zeroing gradients for uninformative negative samples.
  • Achieves superior ranking accuracy over existing cold-start methods and improves item outcome equity.

Why it matters

Cold-start is a major challenge in recommender systems. This paper offers a novel, purely content-based approach that not only improves prediction accuracy but also addresses fairness concerns by promoting equity in item outcomes. Its sparse contrastive learning method is a significant advancement.

Original Abstract

Item cold-start is a pervasive challenge for collaborative filtering (CF) recommender systems. Existing methods often train cold-start models by mapping auxiliary item content, such as images or text descriptions, into the embedding space of a CF model. However, such approaches can be limited by the fundamental information gap between CF signals and content features. In this work, we propose to avoid this limitation with purely content-based modeling of cold items, i.e. without alignment with CF user or item embeddings. We instead frame cold-start prediction in terms of item-item similarity, training a content encoder to project into a latent space where similarity correlates with user preferences. We define our training objective as a sparse generalization of sampled softmax loss with the $α$-entmax family of activation functions, which allows for sharper estimation of item relevance by zeroing gradients for uninformative negatives. We then describe how this Sampled Entmax for Cold-start (SEMCo) training regime can be extended via knowledge distillation, and show that it outperforms existing cold-start methods and standard sampled softmax in ranking accuracy. We also discuss the advantages of purely content-based modeling, particularly in terms of equity of item outcomes.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.