TY - GEN
T1 - Wireless Model Splitting for Communication-Efficient Personalized Federated Learning with Pipeline Parallelism
AU - Wang, Luya
AU - Dong, Yanjie
AU - Zhang, Lei
AU - Hu, Xiping
AU - Cui, Laizhong
AU - Leung, Victor C.M.
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - A wireless federated learning (WFL) system has limited bandwidth and computational power that confines the scale of a supervised learning model. To improve the scalability of the WFL model, a splitting approach has been used to partition the entire model into several sub-models and to assign the sub-models to the server and workers in the WFL system. The previous splitting approaches require to maintain a sub-model for each worker at the server (i.e., parallel splitting) or to sequentially pass the local training information of workers through the sub-model of the server (i.e., sequential splitting). However, when the number of workers is large, the parallel and sequential splitting respectively consume a significant amount of memory and training duration. In this work, we propose a split representation learning (SplitREP) that allows the server and the workers to respectively own the public and private sub-models. Compared with the parallel and sequential splitting, the SplitREP leverages the pipeline parallelism mechanism to reduce the required memory and training duration. Numerical results show that the proposed SplitREP outperforms the benchmarks in the WFL system.
AB - A wireless federated learning (WFL) system has limited bandwidth and computational power that confines the scale of a supervised learning model. To improve the scalability of the WFL model, a splitting approach has been used to partition the entire model into several sub-models and to assign the sub-models to the server and workers in the WFL system. The previous splitting approaches require to maintain a sub-model for each worker at the server (i.e., parallel splitting) or to sequentially pass the local training information of workers through the sub-model of the server (i.e., sequential splitting). However, when the number of workers is large, the parallel and sequential splitting respectively consume a significant amount of memory and training duration. In this work, we propose a split representation learning (SplitREP) that allows the server and the workers to respectively own the public and private sub-models. Compared with the parallel and sequential splitting, the SplitREP leverages the pipeline parallelism mechanism to reduce the required memory and training duration. Numerical results show that the proposed SplitREP outperforms the benchmarks in the WFL system.
KW - Personalized federated learning
KW - pipeline parallelism
KW - wireless model splitting
UR - http://www.scopus.com/inward/record.url?scp=85178612592&partnerID=8YFLogxK
U2 - 10.1109/SPAWC53906.2023.10304511
DO - 10.1109/SPAWC53906.2023.10304511
M3 - Conference contribution
AN - SCOPUS:85178612592
T3 - IEEE Workshop on Signal Processing Advances in Wireless Communications, SPAWC
SP - 421
EP - 425
BT - 2023 IEEE 24th International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2023 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 24th IEEE International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2023
Y2 - 25 September 2023 through 28 September 2023
ER -