Beyond Loss Values: Robust Dynamic Pruning via Loss Trajectory Alignment
Huaiyuan Qin, Muli Yang, Gabriel James Goenawan, Kai Wang, Zheng Wang + 3 more
TLDR
AlignPrune robustly prunes data under noisy labels by using loss trajectory alignment, outperforming existing dynamic pruning methods.
Key contributions
- Proposes AlignPrune, a noise-robust module for dynamic data pruning under label noise.
- Introduces Dynamic Alignment Score (DAS), a loss-trajectory-based criterion for identifying noisy samples.
- DAS improves pruning effectiveness by accurately identifying noisy samples, boosting reliability.
- A plug-and-play module that integrates seamlessly and boosts accuracy by up to 6.3% over baselines.
Why it matters
Existing dynamic pruning struggles with noisy labels, mistakenly keeping bad data. AlignPrune provides a robust, generalizable solution by using loss trajectories, significantly boosting accuracy and enabling better learning in real-world, noisy datasets.
Original Abstract
Existing dynamic data pruning methods often fail under noisy-label settings, as they typically rely on per-sample loss as the ranking criterion. This could mistakenly lead to preserving noisy samples due to their high loss values, resulting in significant performance drop. To address this, we propose AlignPrune, a noise-robust module designed to enhance the reliability of dynamic pruning under label noise. Specifically, AlignPrune introduces the Dynamic Alignment Score (DAS), which is a loss-trajectory-based criterion that enables more accurate identification of noisy samples, thereby improving pruning effectiveness. As a simple yet effective plug-and-play module, AlignPrune can be seamlessly integrated into state-of-the-art dynamic pruning frameworks, consistently outperforming them without modifying either the model architecture or the training pipeline. Extensive experiments on five widely-used benchmarks across various noise types and pruning ratios demonstrate the effectiveness of AlignPrune, boosting accuracy by up to 6.3\% over state-of-the-art baselines. Our results offer a generalizable solution for pruning under noisy data, encouraging further exploration of learning in real-world scenarios. Code is available at: https://github.com/leonqin430/AlignPrune.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.