HFML: heterogeneous hierarchical federated mutual learning on non-IID data

Yang Li, Jie Li, Kan Li*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

3 Citations (Scopus)

Abstract

Non-independent and identical distribution (Non-IID) data and model heterogeneity pose a great challenge for federated learning in cloud-based and edge-based systems. They are easy to lead to inconsistency of gradient updates during the training stage and mismatch of gradient dimensions during the aggregation stage, resulting in the degradation of the global model performance and the consumption of a lot of training time. To solve these problems, this paper proposes a Heterogeneous Hierarchical Federated Mutual Learning (HFML) method in an edge-based system. We design a model assignment mechanism in which clients and edge servers individually fork global models of different structures, and the untrained local models learn mutually with the edge models in deep mutual learning. We use partial periodic aggregation to approximate global aggregation to achieve fast convergence. Our experiments show that HFML obtains state-of-the-art performance than three approaches on common datasets like CIFAR-10/100. Our method improves accuracy up to 2.9% and reduces training time by 30% under homogeneous and heterogeneous models.

Original languageEnglish
JournalAnnals of Operations Research
DOIs
Publication statusAccepted/In press - 2023

Keywords

  • Deep mutual learning
  • Federated learning
  • Heterogeneous models
  • Non-independent and identical distribution (Non-IID)

Fingerprint

Dive into the research topics of 'HFML: heterogeneous hierarchical federated mutual learning on non-IID data'. Together they form a unique fingerprint.

Cite this