DPM++: Dynamic Masked Metric Learning for Occluded Person Re-identification
Lei Tan, Yingshi Luan, Pincong Zou, Pingyang Dai, Liujuan Cao
TLDR
DPM++ introduces dynamic masked metric learning with CLIP-based supervision and saliency-guided data augmentation for robust occluded person re-identification.
Key contributions
- Proposes DPM++, a dynamic masked metric learning framework for robust occluded person re-identification.
- Utilizes CLIP-based two-stage supervision to transfer ID-level semantic priors for masked matching.
- Introduces saliency-guided patch transfer to synthesize realistic occluded samples, improving training.
- Employs occlusion-aware sample pairing and mask-guided optimization for enhanced stability.
Why it matters
Occlusion severely hinders person re-identification in real-world scenarios. This paper presents a unified framework, DPM++, that dynamically adapts to occluded instances, significantly improving re-ID accuracy. Its novel data augmentation and supervision strategies make it highly robust and practical.
Original Abstract
Although person re-identification has made impressive progress, occlusion caused by obstacles remains an unsettled issue in real applications. The difficulty lies in the mismatch between incomplete occluded samples and holistic identity representations. Severe occlusion removes discriminative body cues and introduces interference from background clutter and occluders, making global metric learning unreliable. Existing methods mainly rely on extra pre-trained models to estimate visible parts for alignment or construct occluded samples via data augmentation, but still lack a unified framework that learns robust visibility-consistent matching under realistic occlusion patterns. In this paper, we propose DPM++, a Dynamic Masked Metric Learning framework for occluded person re-identification. DPM++ learns an input-adaptive masked metric that dynamically selects reliable identity subspaces for each occluded instance, enabling matching to emphasize visibility-consistent evidence while suppressing unreliable components. Built upon the classifier-prototype space, DPM++ introduces a CLIP-based two-stage supervision scheme, where ID-level semantic priors are learned from the text branch and transferred into the classifier-prototype space for dynamic masked matching. To strengthen the masked metric, we introduce a saliency-guided patch transfer strategy to synthesize controllable and photo-realistic occluded samples during training. Exploiting real scene priors, this strategy exposes the model to realistic partial observations and provides richer supervision than random erasing. In addition, occlusion-aware sample pairing and mask-guided optimization improve the stability and effectiveness of the framework. Experiments on occluded and holistic person re-identification benchmarks show that DPM++ consistently outperforms previous state-of-the-art methods in both holistic and occlusion scenarios.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.