Beyond One-Size-Fits-All: Adaptive Test-Time Augmentation for Sequential Recommendation
TLDR
AdaTTA uses reinforcement learning to adaptively select test-time augmentation operators for sequential recommendation, improving accuracy over fixed strategies.
Key contributions
- Identifies that "one-size-fits-all" TTA is suboptimal for sequential recommendation due to user heterogeneity.
- Proposes AdaTTA, an RL-based framework that adaptively selects sequence-specific augmentation operators.
- Formulates augmentation selection as an MDP, using an Actor-Critic network with a novel reward design.
- Achieves up to 26.31% relative improvement over fixed TTA strategies on real-world datasets.
Why it matters
This paper addresses a key limitation in Test-Time Augmentation for sequential recommendation by moving beyond uniform strategies. By adaptively selecting augmentations, AdaTTA significantly boosts inference accuracy and mitigates data sparsity without retraining, offering a practical plug-and-play solution.
Original Abstract
Test-time augmentation (TTA) has become a promising approach for mitigating data sparsity in sequential recommendation by improving inference accuracy without requiring costly model retraining. However, existing TTA methods typically rely on uniform, user-agnostic augmentation strategies. We show that this "one-size-fits-all" design is inherently suboptimal, as it neglects substantial behavioral heterogeneity across users, and empirically demonstrate that the optimal augmentation operators vary significantly across user sequences with different characteristics for the first time. To address this limitation, we propose AdaTTA, a plug-and-play reinforcement learning-based adaptive inference framework that learns to select sequence-specific augmentation operators on a per-sequence basis. We formulate augmentation selection as a Markov Decision Process and introduce an Actor-Critic policy network with hybrid state representations and a joint macro-rank reward design to dynamically determine the optimal operator for each input user sequence. Extensive experiments on four real-world datasets and two recommendation backbones demonstrate that AdaTTA consistently outperforms the best fixed-strategy baselines, achieving up to 26.31% relative improvement on the Home dataset while incurring only moderate computational overhead
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.