Is Sliding Window All You Need? An Open Framework for Long-Sequence Recommendation
Sayak Chakrabarty, Souradip Pal
TLDR
An open framework enables practical, efficient long-sequence recommendation training with sliding windows and large vocabularies.
Key contributions
- Provides end-to-end industrial-style long-sequence training with sliding windows.
- Introduces runtime-aware ablation to analyze accuracy-compute trade-offs.
- Proposes k-shift embedding for million-scale vocabularies on commodity GPUs.
- Achieves up to +6.04% MRR and +6.34% Recall@10 with ~4x training overhead.
Why it matters
This paper makes long-sequence recommendation training feasible on modest hardware by open-sourcing a robust pipeline and novel embedding method. It bridges the gap between industrial practice and academic research, enabling broader adoption and experimentation.
Original Abstract
Long interaction histories are central to modern recommender systems, yet training with long sequences is often dismissed as impractical under realistic memory and latency budgets. This work demonstrates that it is not only practical but also effective-at academic scale. We release a complete, end-to-end framework that implements industrial-style long-sequence training with sliding windows, including all data processing, training, and evaluation scripts. Beyond reproducing prior gains, we contribute two capabilities missing from earlier reports: (i) a runtime-aware ablation study that quantifies the accuracy-compute frontier across windowing regimes and strides, and (ii) a novel k-shift embedding layer that enables million-scale vocabularies on commodity GPUs with negligible accuracy loss. Our implementation trains reliably on modest university clusters while delivering competitive retrieval quality (e.g., up to +6.04% MRR and +6.34% Recall@10 on Retailrocket) with $\sim 4 \times $ training-time overheads. By packaging a robust pipeline, reporting training time costs, and introducing an embedding mechanism tailored for low-resource settings, we transform long-sequence training from a closed, industrial technique into a practical, open, and extensible methodology for the community.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.