TY - JOUR
T1 - ISFL
T2 - Federated Learning for Non-i.i.d. Data With Local Importance Sampling
AU - Zhu, Zheqi
AU - Shi, Yuchen
AU - Fan, Pingyi
AU - Peng, Chenghui
AU - Letaief, Khaled B.
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024
Y1 - 2024
N2 - As a promising learning paradigm integrating computation and communication, federated learning (FL) proceeds the local training and the periodic sharing from distributed clients. Due to the non-i.i.d. data distribution on clients, FL model suffers from the gradient diversity, poor performance, bad convergence, etc. In this work, we aim to tackle this key issue by adopting importance sampling (IS) for local training. We propose importance sampling FL(ISFL), an explicit framework with theoretical guarantees. First, we derive the convergence theorem of ISFL to involve the effects of local IS. Then, we formulate the problem of selecting optimal IS weights and obtain the theoretical solutions. We also employ a water-filling method to calculate the IS weights and develop the ISFL algorithms. The experimental results on CIFAR-10 fit the proposed theorems well and verify that ISFL reaps better performance, convergence, sampling efficiency, as well as explainability on the non-i.i.d. data. To the best of our knowledge, ISFL is the first non-i.i.d. FL solution from the local sampling aspect which exhibits theoretical compatibility with neural network models. Furthermore, as a local sampling approach, ISFL can be easily migrated into the other emerging FL frameworks.
AB - As a promising learning paradigm integrating computation and communication, federated learning (FL) proceeds the local training and the periodic sharing from distributed clients. Due to the non-i.i.d. data distribution on clients, FL model suffers from the gradient diversity, poor performance, bad convergence, etc. In this work, we aim to tackle this key issue by adopting importance sampling (IS) for local training. We propose importance sampling FL(ISFL), an explicit framework with theoretical guarantees. First, we derive the convergence theorem of ISFL to involve the effects of local IS. Then, we formulate the problem of selecting optimal IS weights and obtain the theoretical solutions. We also employ a water-filling method to calculate the IS weights and develop the ISFL algorithms. The experimental results on CIFAR-10 fit the proposed theorems well and verify that ISFL reaps better performance, convergence, sampling efficiency, as well as explainability on the non-i.i.d. data. To the best of our knowledge, ISFL is the first non-i.i.d. FL solution from the local sampling aspect which exhibits theoretical compatibility with neural network models. Furthermore, as a local sampling approach, ISFL can be easily migrated into the other emerging FL frameworks.
KW - Federated learning (FL)
KW - importance sampling (IS)
KW - non i.i.d data
KW - water-filling optimization
UR - https://www.webofscience.com/wos/woscc/full-record/WOS:001291138800065
UR - https://openalex.org/W4396754427
UR - https://www.scopus.com/pages/publications/85192986753
U2 - 10.1109/JIOT.2024.3398398
DO - 10.1109/JIOT.2024.3398398
M3 - Journal Article
SN - 2327-4662
VL - 11
SP - 27448
EP - 27462
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 16
ER -