Abstract
Different from its commonly studied scenario to centrally store clients' data in institutions, which implicitly neglects clients' data privacy, we study cross-silo federated learning in a preferable setting to keep private data on clients, and train the global model with a three-layer structure, where the institutions aggregate model updates from their clients for several rounds before sending their aggregated updates to the central server. In this context, we mathematically prove that the number of clients' local training epochs affects the global model performance and thus propose a new approach, Tempo, to adaptively tune the epoch number of each client through training. The results of our evaluation conducted under real network environments show that Tempo can not only improve training performance in terms of global model accuracy and communication efficiency, but also the elapsed training time.
| Original language | English |
|---|---|
| Title of host publication | 2022 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2022 - Proceedings |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 4358-4362 |
| Number of pages | 5 |
| ISBN (Electronic) | 9781665405409 |
| DOIs | |
| Publication status | Published - 2022 |
| Event | 2022 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2022 - Hybrid, Singapore Duration: 22 May 2022 → 27 May 2022 |
Publication series
| Name | ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings |
|---|---|
| Volume | 2022-May |
| ISSN (Print) | 1520-6149 |
Conference
| Conference | 2022 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2022 |
|---|---|
| Country/Territory | Singapore |
| City | Hybrid |
| Period | 22/05/22 → 27/05/22 |
Bibliographical note
Publisher Copyright:© 2022 IEEE
Keywords
- Federated learning
- cross-silo federated learning
- edge computing
- non-IID data