ArXiv TLDR

Exploring the Limits of End-to-End Feature-Affinity Propagation for Single-Point Supervised Infrared Small Target Detection

🐦 Tweet
2605.00722

Qiancheng Zhou, Wenhua Zhang

cs.CV

TLDR

GSACP is an end-to-end method for single-point supervised IRSTD, using in-batch feature propagation to reduce false alarms.

Key contributions

  • Introduces GSACP, an end-to-end method for single-point supervised IRSTD using in-batch feature propagation.
  • Eliminates external pseudo-label construction loops, simplifying the training process significantly.
  • Formalizes "Self-Referential Propagation Drift" and proposes solutions to mitigate this optimization bottleneck.
  • Achieves a 38% reduction in false-positive artifacts on SIRST3, setting a new ultra-low false-alarm regime.

Why it matters

This paper offers a simpler, end-to-end approach to single-point supervised IRSTD, significantly cutting annotation costs. Its method, GSACP, achieves superior false-alarm suppression, which is critical for real-world deployment scenarios.

Original Abstract

Single-point supervised infrared small target detection (IRSTD) drastically reduces dense annotation costs. Current state-of-the-art (SOTA) methods achieve high precision by recovering mask supervision through explicit, offline pseudo-label construction, such as multi-stage active learning and physics-driven mask generation. In this paper, we study a minimalist alternative: generating point-to-mask supervision online through in-batch, point-anchored feature-affinity propagation. We instantiate this paradigm as GSACP, an end-to-end testbed that directly supervises the detector using hard-margin feature affinity gated by local image priors, entirely eliminating external label-evolution loops. This compact design, however, exposes an optimization bottleneck. Because the affinity target is generated from the same feature representation being optimized, training forms a self-referential loop. We theoretically formalize this as \emph{Self-Referential Propagation Drift}, a representation-supervision entanglement that can sharpen true boundaries or distort the feature space to satisfy its own targets. To systematically isolate these failure modes, we apply a protocolized single-variable ablation procedure spanning local EMA teacher decoupling, hard-background contrastive separation, and adaptive support geometry. On the SIRST3 dataset, GSACP-Final establishes a new ultra-low false-alarm operating regime, achieving a highly competitive $0.6674$ mIoU while demonstrating a $38\% relative reduction in false-positive artifacts ($\mathrm{Fa}$) compared with PAL. By systematically deconstructing the end-to-end paradigm, we map its performance boundaries and show that in-batch feature propagation provides a compact alternative for deployment scenarios where false-alarm suppression is paramount.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.