Abstract
To facilitate Federated Learning (FL) on low-memory embedded devices, various federated pruning methods aim to reduce memory usage during inference but have a limited impact on training memory burdens. Alternatively, zeroth-order or backpropagation-free (BP-Free) methods can partially alleviate memory consumption but still face computational overhead as the number of model parameters increases. To address these issues, we propose a memory-efficient federated foresight pruning method based on the Neural Tangent Kernel (NTK), which seamlessly integrates with federated BP-Free training frameworks. We approximate federated NTK using local NTK matrices and demonstrate that the data-free property of our method significantly reduces approximation error in highly heterogeneous data scenarios. Our method improves the vanilla BP-Free method with fewer floating point operations (FLOPs) and alleviates memory pressure during pruning and training, making FL more feasible for low-memory devices. Experimental results on simulation- and real test-bed-based platforms show that our method improves the accuracy by up to 6.35% and reduces the FLOPs by up to 57% against the vanilla BP-Free method while maintaining the same 9× memory usage saving.
| Original language | English |
|---|---|
| Article number | 103526 |
| Journal | Journal of Systems Architecture |
| Volume | 168 |
| DOIs | |
| State | Published - Nov 2025 |
Keywords
- Federated learning
- Memory efficiency
- Model pruning
- Zeroth-order optimization