ArXiv TLDR

FedSIR: Spectral Client Identification and Relabeling for Federated Learning with Noisy Labels

🐦 Tweet
2604.20825

Sina Gholami, Abdulmoneam Ali, Tania Haghighi, Ahmed Arafa, Minhaj Nur Alam

cs.LGcs.AIcs.CVcs.DCeess.SP

TLDR

FedSIR is a federated learning framework that uses spectral analysis to identify noisy clients and relabel corrupted samples, improving robustness.

Key contributions

  • Identifies clean and noisy clients by analyzing spectral consistency of class-wise feature subspaces.
  • Enables noisy clients to relabel corrupted samples using spectral references from clean clients.
  • Employs a noise-aware training strategy with logit-adjusted loss, distillation, and distance-aware aggregation.

Why it matters

Federated learning often suffers from noisy labels, degrading model performance. FedSIR introduces a novel spectral approach to identify and mitigate this noise, outperforming existing methods. This significantly enhances the robustness and reliability of FL systems.

Original Abstract

Federated learning (FL) enables collaborative model training without sharing raw data; however, the presence of noisy labels across distributed clients can severely degrade the learning performance. In this paper, we propose FedSIR, a multi-stage framework for robust FL under noisy labels. Different from existing approaches that mainly rely on designing noise-tolerant loss functions or exploiting loss dynamics during training, our method leverages the spectral structure of client feature representations to identify and mitigate label noise. Our framework consists of three key components. First, we identify clean and noisy clients by analyzing the spectral consistency of class-wise feature subspaces with minimal communication overhead. Second, clean clients provide spectral references that enable noisy clients to relabel potentially corrupted samples using both dominant class directions and residual subspaces. Third, we employ a noise-aware training strategy that integrates logit-adjusted loss, knowledge distillation, and distance-aware aggregation to further stabilize federated optimization. Extensive experiments on standard FL benchmarks demonstrate that FedSIR consistently outperforms state-of-the-art methods for FL with noisy labels. The code is available at https://github.com/sinagh72/FedSIR.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.