TY - JOUR
T1 - GossipFL
T2 - A Decentralized Federated Learning Framework With Sparsified and Adaptive Communication
AU - Tang, Zhenheng
AU - Shi, Shaohuai
AU - Li, Bo
AU - Chu, Xiaowen
N1 - Funding information:
The research was supported in part by the RGC RIF under Grant R6021-20, and in part by RGC GRF under Grants 16209120 and 16200221. (Corresponding authors: Shaohuai Shi and Xiaowen Chu.)
Publisher copyright:
© 2022 IEEE.
PY - 2023/3/1
Y1 - 2023/3/1
N2 - Recently, federated learning (FL) techniques have enabled multiple users to train machine learning models collaboratively without data sharing. However, existing FL algorithms suffer from the communication bottleneck due to network bandwidth pressure and/or low bandwidth utilization of the participating clients in both centralized and decentralized architectures. To deal with the communication problem while preserving the convergence performance, we introduce a communication-efficient decentralized FL framework GossipFL. In GossipFL, we 1) design a novel sparsification algorithm to enable that each client only needs to communicate with one peer with a highly sparsified model, and 2) propose a new and novel gossip matrix generation algorithm that can better utilize the bandwidth resources while preserving the convergence property. We also theoretically prove that GossipFL has convergence guarantees. We conduct experiments with three convolutional neural networks on two datasets (IID and non-IID) under two distributed environments (14 clients and 100 clients) to verify the effectiveness of GossipFL. Experimental results show that GossipFL takes less communication traffic for 38.5% and less communication time for 49.8% than state-of-the-art solutions while achieving comparative model accuracy.
AB - Recently, federated learning (FL) techniques have enabled multiple users to train machine learning models collaboratively without data sharing. However, existing FL algorithms suffer from the communication bottleneck due to network bandwidth pressure and/or low bandwidth utilization of the participating clients in both centralized and decentralized architectures. To deal with the communication problem while preserving the convergence performance, we introduce a communication-efficient decentralized FL framework GossipFL. In GossipFL, we 1) design a novel sparsification algorithm to enable that each client only needs to communicate with one peer with a highly sparsified model, and 2) propose a new and novel gossip matrix generation algorithm that can better utilize the bandwidth resources while preserving the convergence property. We also theoretically prove that GossipFL has convergence guarantees. We conduct experiments with three convolutional neural networks on two datasets (IID and non-IID) under two distributed environments (14 clients and 100 clients) to verify the effectiveness of GossipFL. Experimental results show that GossipFL takes less communication traffic for 38.5% and less communication time for 49.8% than state-of-the-art solutions while achieving comparative model accuracy.
KW - Deep learning
KW - communication efficiency
KW - federated learning
UR - http://www.scopus.com/inward/record.url?scp=85146237939&partnerID=8YFLogxK
U2 - 10.1109/TPDS.2022.3230938
DO - 10.1109/TPDS.2022.3230938
M3 - Journal article
SN - 1045-9219
VL - 34
SP - 909
EP - 922
JO - IEEE Transactions on Parallel and Distributed Systems
JF - IEEE Transactions on Parallel and Distributed Systems
IS - 3
ER -