FedKL: Tackling Data Heterogeneity in Federated Reinforcement Learning by Penalizing KL Divergence

Zhijie XIE, Shenghui Song*

*Corresponding author for this work

Research output: Contribution to journalJournal Articlepeer-review

56 Citations (Scopus)

Abstract

One of the fundamental issues for Federated Learning (FL) is data heterogeneity, which causes accuracy degradation, slow convergence, and the communication bottleneck issue. Although the impact of data heterogeneity on supervised FL has been widely studied, the related investigation for Federated Reinforcement Learning (FRL) is still in its infancy. In this paper, we first define the type and level of data heterogeneity for FRL systems. By inspecting the connection between the global and local objective functions, we prove that local training can benefit the global objective, if the local update is properly penalized by the total variation (TV) distance between the local and global policies. A necessary condition for the global policy to be learn-able from the local environments is also derived, which is directly related to the heterogeneity level. Based on the theoretical result, a Kullback-Leibler (KL) divergence based penalty is proposed to directly constrain the model outputs in the distribution space and the convergence proof of the proposed algorithm is also provided. By jointly penalizing the divergence of the local policy from the global policy with a global penalty and penalizing each iteration of the local training with a local penalty, the proposed method achieves a better trade-off between training speed (step size) and convergence. Experiment results on two popular Reinforcement Learning (RL) experiment platforms demonstrate the advantage of the proposed algorithm over existing methods in accelerating and stabilizing the training process with heterogeneous data.

Original languageEnglish
Pages (from-to)1227-1242
Number of pages16
JournalIEEE Journal on Selected Areas in Communications
Volume41
Issue number4
Early online date6 Feb 2023
DOIs
Publication statusPublished - 1 Apr 2023

Bibliographical note

Publisher Copyright:
© 1983-2012 IEEE.

Keywords

  • Federated reinforcement learning
  • data heterogeneity
  • policy gradient

Fingerprint

Dive into the research topics of 'FedKL: Tackling Data Heterogeneity in Federated Reinforcement Learning by Penalizing KL Divergence'. Together they form a unique fingerprint.

Cite this