TY - GEN
T1 - DFDG
T2 - 24th IEEE International Conference on Data Mining, ICDM 2024
AU - Luo, Kangyang
AU - Wang, Shuai
AU - Fu, Yexuan
AU - Shao, Renrong
AU - Li, Xiang
AU - Lan, Yunshi
AU - Gao, Ming
AU - Shu, Jinlong
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Federated Learning (FL) is a distributed machine learning scheme in which clients jointly participate in the collaborative training of a global model by sharing model information rather than their private datasets. In light of concerns associated with communication and privacy, one-shot FL with a single communication round has emerged as a de facto promising solution. However, existing one-shot FL methods either require public datasets, focus on model homogeneous settings, or distill limited knowledge from local models, making it difficult or even impractical to train a robust global model. To address these limitations, we propose a new data-free dual-generator adversarial distillation method (namely DFDG) for one-shot FL, which can explore a broader local models' training space via training dual generators. DFDG is executed in an adversarial manner and comprises two parts: dual-generator training and dual-model distillation. In dual-generator training, we delve into each generator concerning fidelity, transferability and diversity to ensure its utility, and additionally tailor the cross-divergence loss to lessen the overlap of dual generators' output spaces. In dual-model distillation, the trained dual generators work together to provide the training data for updates of the global model. At last, our extensive experiments on various image classification tasks show that DFDG achieves significant performance gains in accuracy compared to SOTA baselines. We provide our code here: https://anonymous.4open.science/r/DFDG-7BDB.
AB - Federated Learning (FL) is a distributed machine learning scheme in which clients jointly participate in the collaborative training of a global model by sharing model information rather than their private datasets. In light of concerns associated with communication and privacy, one-shot FL with a single communication round has emerged as a de facto promising solution. However, existing one-shot FL methods either require public datasets, focus on model homogeneous settings, or distill limited knowledge from local models, making it difficult or even impractical to train a robust global model. To address these limitations, we propose a new data-free dual-generator adversarial distillation method (namely DFDG) for one-shot FL, which can explore a broader local models' training space via training dual generators. DFDG is executed in an adversarial manner and comprises two parts: dual-generator training and dual-model distillation. In dual-generator training, we delve into each generator concerning fidelity, transferability and diversity to ensure its utility, and additionally tailor the cross-divergence loss to lessen the overlap of dual generators' output spaces. In dual-model distillation, the trained dual generators work together to provide the training data for updates of the global model. At last, our extensive experiments on various image classification tasks show that DFDG achieves significant performance gains in accuracy compared to SOTA baselines. We provide our code here: https://anonymous.4open.science/r/DFDG-7BDB.
KW - Data heterogeneity
KW - Data-free knowledge distillation
KW - Model heterogeneity
KW - One-shot Federated Learning
UR - https://www.scopus.com/pages/publications/86000225749
U2 - 10.1109/ICDM59182.2024.00035
DO - 10.1109/ICDM59182.2024.00035
M3 - 会议稿件
AN - SCOPUS:86000225749
T3 - Proceedings - IEEE International Conference on Data Mining, ICDM
SP - 281
EP - 290
BT - Proceedings - 24th IEEE International Conference on Data Mining, ICDM 2024
A2 - Baralis, Elena
A2 - Zhang, Kun
A2 - Damiani, Ernesto
A2 - Debbah, Merouane
A2 - Kalnis, Panos
A2 - Wu, Xindong
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 9 December 2024 through 12 December 2024
ER -