TY - JOUR
T1 - AsyFed
T2 - Accelerated Federated Learning with Asynchronous Communication Mechanism
AU - Li, Zhixin
AU - Huang, Chunpu
AU - Gai, Keke
AU - Lu, Zhihui
AU - Wu, Jie
AU - Chen, Lulu
AU - Xu, Yangchuan
AU - Choo, Kim Kwang Raymond
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2023/5/15
Y1 - 2023/5/15
N2 - As a new distributed machine learning (ML) framework for privacy protection, federated learning (FL) enables substantial Internet of Things (IoT) devices (e.g., mobile phones, tablets, etc.) to participate in collaborative training of an ML model. FL can protect the data privacy of IoT devices without exposing their raw data. However, the diversity of IoT devices may degrade the overall training process due to the straggler issue. To tackle this problem, we propose a gear-based asynchronous FL (AsyFed) architecture. It adds a gear layer between the clients and the FL server as a mediator to store the model parameters. The key insight is that we group these clients with similar training abilities into the same gear. The clients within the same gear conduct synchronous training. These gears then communicate with the global FL server asynchronously. Besides, we propose a T-step mechanism to reduce the weight from the slow gear when they are communicating with the FL server. The extensive experiment evaluations indicate that AsyFed outperforms FedAvg (baseline synchronous FL scheme) and some state-of-the-art asynchronous FL methods in terms of training accuracy or speed under different data distributions. The only negligible overhead is that we leverage the extra layer (gear layer) to preserve part of the model parameters.
AB - As a new distributed machine learning (ML) framework for privacy protection, federated learning (FL) enables substantial Internet of Things (IoT) devices (e.g., mobile phones, tablets, etc.) to participate in collaborative training of an ML model. FL can protect the data privacy of IoT devices without exposing their raw data. However, the diversity of IoT devices may degrade the overall training process due to the straggler issue. To tackle this problem, we propose a gear-based asynchronous FL (AsyFed) architecture. It adds a gear layer between the clients and the FL server as a mediator to store the model parameters. The key insight is that we group these clients with similar training abilities into the same gear. The clients within the same gear conduct synchronous training. These gears then communicate with the global FL server asynchronously. Besides, we propose a T-step mechanism to reduce the weight from the slow gear when they are communicating with the FL server. The extensive experiment evaluations indicate that AsyFed outperforms FedAvg (baseline synchronous FL scheme) and some state-of-the-art asynchronous FL methods in terms of training accuracy or speed under different data distributions. The only negligible overhead is that we leverage the extra layer (gear layer) to preserve part of the model parameters.
KW - Federated learning (FL)
KW - asynchronous update
KW - communication overhead
KW - device heterogeneity
UR - http://www.scopus.com/inward/record.url?scp=85146217231&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2022.3231913
DO - 10.1109/JIOT.2022.3231913
M3 - Article
AN - SCOPUS:85146217231
SN - 2327-4662
VL - 10
SP - 8670
EP - 8683
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 10
ER -