TY - JOUR
T1 - POP-FL
T2 - Towards Efficient Federated Learning on Edge Using Parallel Over-Parameterization
AU - Lu, Xingjian
AU - Zheng, Haikun
AU - Liu, Wenyan
AU - Jiang, Yuhui
AU - Wu, Hongyue
N1 - Publisher Copyright:
© 2008-2012 IEEE.
PY - 2024/3/1
Y1 - 2024/3/1
N2 - Federated Learning (FL) is a promising paradigm for mining massive data while respecting users' privacy. However, the deployment of FL on resource-constrained edge devices remains elusive due to its high resource demand. In this paper, unlike existing works that use expensive dense models, we propose to utilize dynamic sparse training in FL and design a novel sparse-to-sparse FL framework, named as POP-FL. The framework can reduce both computation and communication overheads while maintaining the performance of the global model. Specifically, POP-FL partitions massive clients into groups and performs parallel parameter exploration, i.e., Parallel Over-Parameterization, over the collaboration between these groups. This exploration can greatly improve the expressibility and generalizability of sparse training in FL (especially for extreme sparsity levels) through reliably covering sufficient parameters and dynamically updating the global sparse network's structure during the training process. Experimental results show that compared with existing sparse-to-sparse training methods in both iid and non-iid data distribution, POP-FL achieves the best inference accuracy on various representative networks.
AB - Federated Learning (FL) is a promising paradigm for mining massive data while respecting users' privacy. However, the deployment of FL on resource-constrained edge devices remains elusive due to its high resource demand. In this paper, unlike existing works that use expensive dense models, we propose to utilize dynamic sparse training in FL and design a novel sparse-to-sparse FL framework, named as POP-FL. The framework can reduce both computation and communication overheads while maintaining the performance of the global model. Specifically, POP-FL partitions massive clients into groups and performs parallel parameter exploration, i.e., Parallel Over-Parameterization, over the collaboration between these groups. This exploration can greatly improve the expressibility and generalizability of sparse training in FL (especially for extreme sparsity levels) through reliably covering sufficient parameters and dynamically updating the global sparse network's structure during the training process. Experimental results show that compared with existing sparse-to-sparse training methods in both iid and non-iid data distribution, POP-FL achieves the best inference accuracy on various representative networks.
KW - Distributed machine learning
KW - edge computing
KW - federated learning
KW - model sparse training
UR - https://www.scopus.com/pages/publications/85188002697
U2 - 10.1109/TSC.2024.3376194
DO - 10.1109/TSC.2024.3376194
M3 - 文章
AN - SCOPUS:85188002697
SN - 1939-1374
VL - 17
SP - 617
EP - 630
JO - IEEE Transactions on Services Computing
JF - IEEE Transactions on Services Computing
IS - 2
ER -