A Hierarchical Sampling Framework for bounding the Generalization Error of Federated Learning
Dario Filatrella, Ragnar Thobaben, Mikael Skoglund
TLDR
This paper introduces a hierarchical sampling framework to derive tight generalization bounds for Federated Learning using Wasserstein distance.
Key contributions
- Introduces a hierarchical sampling framework for Federated Learning (FL) with a multi-layered tree structure.
- Derives generalization bounds using Wasserstein distance under Lipschitz loss via a supersample construction.
- Recovers and strictly improves existing conditional mutual information (CMI) bounds for bounded losses.
- Demonstrates applicability with Differential Privacy and achieves asymptotic rate in Gaussian Location Model.
Why it matters
This paper provides a novel, tighter framework for understanding generalization in Hierarchical Federated Learning. It advances the theoretical understanding of FL's reliability, especially in complex data dependencies, and integrates with privacy assumptions.
Original Abstract
We study expected generalization bounds for the Hierarchical Federated Learning (HFL) setup using Wasserstein distance. We introduce a generalized framework in which data is sampled hierarchically, and we model it with a multi-layered tree structure that induces dependencies among the clients' datasets. We derive generalization bounds in terms of Wasserstein distance under the Lipschitz assumption on the loss function, by applying a supersample construction that allows us to measure the sensitivity of the algorithm to the change of a single node in the sampling tree. By leveraging the FL structure, we recover and strictly imply existing state-of-the-art conditional mutual information (CMI) bounds in the case of bounded losses. We also show that our bound can be applied together with Differential Privacy assumptions, to recover generalization bounds based on algorithmic privacy. To assess the tightness of our bounds, we study the Gaussian Location Model (GLM) and show that we recover the actual asymptotic rate of the generalization error.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.