Federated Learning under Distributed Concept Drift
- Ellango Jothimurugesan ,
- Kevin Hsieh ,
- Jianyu Wang ,
- Gauri Joshi ,
- Phillip B. Gibbons
International Conference on Artificial Intelligence and Statistics (AISTATS) |
Selected as Oral Presentation (Top 2% Submissions)
Federated Learning (FL) under distributed concept drift is a largely unexplored area. Although concept drift is itself a well-studied phenomenon, it poses particular challenges for FL, because drifts arise staggered in time and space (across clients). To the best of our knowledge, this work is the first to explicitly study data heterogeneity in both dimensions. We first demonstrate that prior solutions to drift adaptation that use a single global model are ill-suited to staggered drifts, necessitating multiple-model solutions. We identify the problem of drift adaptation as a time-varying clustering problem, and we propose two new clustering algorithms for reacting to drifts based on local drift detection and hierarchical clustering. Empirical evaluation shows that our solutions achieve significantly higher accuracy than existing baselines, and are comparable to an idealized algorithm with oracle knowledge of the ground-truth clustering of clients to concepts at each time step.
Publication Downloads
Federated Learning under Distributed Concept Drift (FedDrift)
May 9, 2024
This repository is the source code for our paper: Federated Learning under Distributed Concept Drift (AISTATS'23).