Collaborative Neural Architecture Search for Personalized Federated Learning

Yi Liu, Song Guo, Jie Zhang*, Zicong Hong, Yufeng Zhan, Qihua Zhou

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Personalized federated learning (pFL) is a promising approach to train customized models for multiple clients over heterogeneous data distributions. However, existing works on pFL often rely on the optimization of model parameters and ignore the personalization demand on neural network architecture, which can greatly affect the model performance in practice. Therefore, generating personalized models with different neural architectures for different clients is a key issue in implementing pFL in a heterogeneous environment. Motivated by Neural Architecture Search (NAS), a model architecture searching methodology, this paper aims to automate the model design in a collaborative manner while achieving good training performance for each client. Specifically, we reconstruct the centralized searching of NAS into the distributed scheme called Personalized Architecture Search (PAS), where differentiable architecture fine-tuning is achieved via gradient-descent optimization, thus making each client obtain the most appropriate model. Furthermore, to aggregate knowledge from heterogeneous neural architectures, a knowledge distillation-based training framework is proposed to achieve a good trade-off between generalization and personalization in federated learning. Extensive experiments demonstrate that our architecture-level personalization method achieves higher accuracy under the non-iid settings, while not aggravating model complexity over state-of-the-art benchmarks.

源语言英语
期刊IEEE Transactions on Computers
DOI
出版状态已接受/待刊 - 2024

指纹

探究 'Collaborative Neural Architecture Search for Personalized Federated Learning' 的科研主题。它们共同构成独一无二的指纹。

引用此