ArXiv TLDR

FuTCR: Future-Targeted Contrast and Repulsion for Continual Panoptic Segmentation

🐦 Tweet
2605.12451

Nicholas Ikechukwu, Keanu Nichols, Deepti Ghadiyaram, Bryan A. Plummer

cs.CV

TLDR

FuTCR improves continual panoptic segmentation by pre-structuring representations for future classes, boosting new-class performance.

Key contributions

  • Addresses CPS limitation where unlabeled pixels are grouped as "background," hindering new class learning.
  • Discovers "future-like" regions from unlabeled pixels by analyzing non-background logits.
  • Applies pixel-to-region contrast to build prototypes for future categories and repels background features.
  • Achieves up to 28% relative new-class PQ improvement and 4% base-class gain over state-of-the-art.

Why it matters

Current CPS methods struggle with new categories because they treat all unlabeled pixels as generic background. This paper introduces a novel approach to proactively prepare the model for future classes. By reserving representational space, it significantly improves adaptation to new objects without sacrificing existing knowledge, which is crucial for real-world applications.

Original Abstract

Continual Panoptic Segmentation (CPS) requires methods that can quickly adapt to new categories over time. The nature of this dense prediction task means that training images may contain a mix of labeled and unlabeled objects. As nothing is known about these unlabeled objects a priori, existing methods often simply group any unlabeled pixel into a single "background" class during training. In effect, during training, they repeatedly tell the model that all the different background categories are the same (even when they aren't). This makes learning to identify different background categories as they are added challenging since these new categories may require using information the model was previously told was unimportant and ignored. Thus, we propose a Future-Targeted Contrastive and Repulsive (FuTCR) framework that addresses this limitation by restructuring representations before new classes are introduced. FuTCR first discovers confident future-like regions by grouping model-predicted masks whose pixels are consistently classified as background but exhibit non-background logits. Next, FuTCR applies pixel-to-region contrast to build coherent prototypes from these unlabeled regions, while simultaneously repelling background features away from known-class prototypes to explicitly reserve representational space for future categories. Experiments across six CPS settings and a range of dataset sizes show FuTCR improves relative new-class panoptic quality over the state-of-the-art by up to 28%, while preserving or improving base-class performance with gains up to 4%.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.