TY - GEN
T1 - Enhancing Model Performance via Vertical Federated Learning for Non-Overlapping Data Utilization
AU - Wu, Bing
AU - Dong, Xiaolei
AU - Shen, Jiachen
AU - Cao, Zhenfu
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Collaborative training of machine learning models is essential in the era of big data. Federated learning ensures secure data sharing among multiple parties without compromising privacy. It includes various approaches like horizontal federated learning, vertical federated learning, and federated transfer learning. Vertical federated learning enables participants to train on different feature spaces while sharing sample labels. However, existing vertical federated learning schemes rely on participants having sufficient overlapping samples, limiting their effectiveness in scenarios with limited overlapping data. This poses challenges, particularly in domains like the medical industry where collecting enough overlapping samples is difficult. Traditional approaches fail to utilize the non-overlapping portion of the sample data, resulting in suboptimal model performance due to insufficient training data. To address this issue, we propose a novel scheme for training neural network models within the vertical federated learning framework using non-overlapping samples. Our scheme leverages fuzzy prediction to handle non-overlapping samples, improving data utilization and enhancing model performance. Crucially, our approach ensures participants' data privacy by not requiring the sharing of original data or model parameters. Experimental results validate the efficacy and efficiency of our proposed scheme.
AB - Collaborative training of machine learning models is essential in the era of big data. Federated learning ensures secure data sharing among multiple parties without compromising privacy. It includes various approaches like horizontal federated learning, vertical federated learning, and federated transfer learning. Vertical federated learning enables participants to train on different feature spaces while sharing sample labels. However, existing vertical federated learning schemes rely on participants having sufficient overlapping samples, limiting their effectiveness in scenarios with limited overlapping data. This poses challenges, particularly in domains like the medical industry where collecting enough overlapping samples is difficult. Traditional approaches fail to utilize the non-overlapping portion of the sample data, resulting in suboptimal model performance due to insufficient training data. To address this issue, we propose a novel scheme for training neural network models within the vertical federated learning framework using non-overlapping samples. Our scheme leverages fuzzy prediction to handle non-overlapping samples, improving data utilization and enhancing model performance. Crucially, our approach ensures participants' data privacy by not requiring the sharing of original data or model parameters. Experimental results validate the efficacy and efficiency of our proposed scheme.
KW - Data privacy
KW - Neural network
KW - Non-overlapping data
KW - Vertical federated learning
UR - https://www.scopus.com/pages/publications/85172871054
U2 - 10.1109/ISPDS58840.2023.10235507
DO - 10.1109/ISPDS58840.2023.10235507
M3 - 会议稿件
AN - SCOPUS:85172871054
T3 - 2023 4th International Conference on Information Science, Parallel and Distributed Systems, ISPDS 2023
SP - 549
EP - 554
BT - 2023 4th International Conference on Information Science, Parallel and Distributed Systems, ISPDS 2023
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 4th International Conference on Information Science, Parallel and Distributed Systems, ISPDS 2023
Y2 - 14 July 2023 through 16 July 2023
ER -