TY - JOUR
T1 - Adaptive Federated Learning on Non-IID Data With Resource Constraint
AU - Zhang, Jie
AU - Guo, Song
AU - Qu, Zhihao
AU - Zeng, Deze
AU - Zhan, Yufeng
AU - Liu, Qifeng
AU - Akerkar, Rajendra
N1 - Publisher Copyright:
© 1968-2012 IEEE.
PY - 2022/7/1
Y1 - 2022/7/1
N2 - Federated learning (FL) has been widely recognized as a promising approach by enabling individual end-devices to cooperatively train a global model without exposing their own data. One of the key challenges in FL is the non-independent and identically distributed (Non-IID) data across the clients, which decreases the efficiency of stochastic gradient descent (SGD) based training process. Moreover, clients with different data distributions may cause bias to the global model update, resulting in a degraded model accuracy. To tackle the Non-IID problem in FL, we aim to optimize the local training process and global aggregation simultaneously. For local training, we analyze the effect of hyperparameters (e.g., the batch size, the number of local updates) on the training performance of FL. Guided by the toy example and theoretical analysis, we are motivated to mitigate the negative impacts incurred by Non-IID data via selecting a subset of participants and adaptively adjust their batch size. A deep reinforcement learning based approach has been proposed to adaptively control the training of local models and the phase of global aggregation. Extensive experiments on different datasets show that our method can improve the model accuracy by up to 30 percent, as compared to the state-of-the-art approaches.
AB - Federated learning (FL) has been widely recognized as a promising approach by enabling individual end-devices to cooperatively train a global model without exposing their own data. One of the key challenges in FL is the non-independent and identically distributed (Non-IID) data across the clients, which decreases the efficiency of stochastic gradient descent (SGD) based training process. Moreover, clients with different data distributions may cause bias to the global model update, resulting in a degraded model accuracy. To tackle the Non-IID problem in FL, we aim to optimize the local training process and global aggregation simultaneously. For local training, we analyze the effect of hyperparameters (e.g., the batch size, the number of local updates) on the training performance of FL. Guided by the toy example and theoretical analysis, we are motivated to mitigate the negative impacts incurred by Non-IID data via selecting a subset of participants and adaptively adjust their batch size. A deep reinforcement learning based approach has been proposed to adaptively control the training of local models and the phase of global aggregation. Extensive experiments on different datasets show that our method can improve the model accuracy by up to 30 percent, as compared to the state-of-the-art approaches.
KW - Federated learning
KW - batch size adaption
KW - deep reinforcement learning
KW - non-IID data
UR - http://www.scopus.com/inward/record.url?scp=85112664813&partnerID=8YFLogxK
U2 - 10.1109/TC.2021.3099723
DO - 10.1109/TC.2021.3099723
M3 - Article
AN - SCOPUS:85112664813
SN - 0018-9340
VL - 71
SP - 1655
EP - 1667
JO - IEEE Transactions on Computers
JF - IEEE Transactions on Computers
IS - 7
ER -