Evaluating Differential Privacy Against Membership Inference in Federated Learning: Insights from the NIST Genomics Red Team Challenge
TLDR
This paper evaluates Differential Privacy (DP) against membership inference attacks (MIAs) in Federated Learning, showing a stacking attack can still leak info.
Key contributions
- Evaluates Differential Privacy (DP) against membership inference attacks (MIAs) in Federated Learning.
- Proposes a novel stacking attack strategy using seven black-box estimators to improve MIA accuracy.
- Shows the stacking attack maintains measurable membership leakage even at low DP (ε=200).
- Provides empirical insights into how stacking-based inference degrades across calibrated DP tiers.
Why it matters
This paper is crucial for understanding the true privacy guarantees of Differential Privacy in Federated Learning. It demonstrates that sophisticated membership inference attacks can still leak information, even with DP applied. These insights are vital for designing more robust privacy-preserving systems.
Original Abstract
While Federated Learning (FL) mitigates direct data exposure, the resulting trained models remain susceptible to membership inference attacks (MIAs). This paper presents an empirical evaluation of Differential Privacy (DP) as a defense mechanism against MIAs in FL, leveraging the environment of the 2025 NIST Genomics Privacy-Preserving Federated Learning (PPFL) Red Teaming Event. To improve inference accuracy, we propose a stacking attack strategy that ensembles seven black-box estimators to train a meta-classifier on prediction probabilities and cross-entropy losses. We evaluate this methodology against target models under three privacy configurations: an unprotected convolutional neural network (CNN, $ε=\infty$), a low-privacy DP model ($ε=200$), and a high-privacy DP model ($ε=10$). The attack outperforms all baselines in the No DP and Low Privacy settings and, critically, maintains measurable membership leakage at $ε=200$ where a single-signal LiRA baseline collapses. Evaluated on an independent third-party benchmark, these results provide an empirical characterisation of how stacking-based inference degrades across calibrated DP tiers in FL.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.