Abstract
Federated Learning has attracted widespread attention in recent years because it allows massive edge nodes to collaboratively train machine learning models without sharing their private data sets. However, these edge nodes are usually heterogeneous in computational capability and statistically different in data distribution, i.e., non-independent and identically distributed (IID), leading to significant performance degradation. Although status quo asynchronous training methods can solve the heterogeneity issue, they cannot prevent the non-IID problem from reducing the convergence rate. In this article, we propose a novel paradigm that schedules the gradient with partially averaged gradients and applies the global momentum (GSGM) for asynchronous training over non-IID data sets in an edge environment. Our key idea is to apply global momentum and partial average on the biased gradients calculated on edge nodes after scheduling, to make the training process stable. Empirical results demonstrate that GSGM can well adapt to different degrees of non-IID data and bring 20% performance gains in terms of training stability for popular optimization algorithms with enhanced accuracy over Fashion-Mnist and CIFAR-10 data sets.
| Original language | English |
|---|---|
| Pages (from-to) | 18817-18828 |
| Number of pages | 12 |
| Journal | IEEE Internet of Things Journal |
| Volume | 9 |
| Issue number | 19 |
| DOIs | |
| Publication status | Published - 1 Oct 2022 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© 2014 IEEE.
Keywords
- Asynchronous
- Federated Learning
- edge computing
- global momentum
- gradient scheduling
- non-independent and identically distributed (IID) data
Fingerprint
Dive into the research topics of 'Gradient Scheduling With Global Momentum for Asynchronous Federated Learning in Edge Environment'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver