arXiv Analytics

Sign in

arXiv:2403.08364 [cs.LG]AbstractReferencesReviewsResources

Decoupled Federated Learning on Long-Tailed and Non-IID data with Feature Statistics

Zhuoxin Chen, Zhenyu Wu, Yang Ji

Published 2024-03-13Version 1

Federated learning is designed to enhance data security and privacy, but faces challenges when dealing with heterogeneous data in long-tailed and non-IID distributions. This paper explores an overlooked scenario where tail classes are sparsely distributed over a few clients, causing the models trained with these classes to have a lower probability of being selected during client aggregation, leading to slower convergence rates and poorer model performance. To address this issue, we propose a two-stage Decoupled Federated learning framework using Feature Statistics (DFL-FS). In the first stage, the server estimates the client's class coverage distributions through masked local feature statistics clustering to select models for aggregation to accelerate convergence and enhance feature learning without privacy leakage. In the second stage, DFL-FS employs federated feature regeneration based on global feature statistics and utilizes resampling and weighted covariance to calibrate the global classifier to enhance the model's adaptability to long-tailed data distributions. We conducted experiments on CIFAR10-LT and CIFAR100-LT datasets with various long-tailed rates. The results demonstrate that our method outperforms state-of-the-art methods in both accuracy and convergence rate.

Related articles: Most relevant | Search more
arXiv:2104.10223 [cs.LG] (Published 2021-04-20)
More Than Meets The Eye: Semi-supervised Learning Under Non-IID Data
arXiv:2409.05347 [cs.LG] (Published 2024-09-09)
TriplePlay: Enhancing Federated Learning with CLIP for Non-IID Data and Resource Efficiency
arXiv:2103.15947 [cs.LG] (Published 2021-03-29)
Federated Learning with Taskonomy for Non-IID Data