Abstract
As an emerging technique, mobile-edge computing (MEC) introduces a new scheme for various distributed communication-computing systems, such as industrial Internet of Things (IoT), vehicular communication, smart city, etc. In this work, we mainly focus on the timeliness of the MEC systems where the freshness of the data and computation tasks is significant. First, we formulate a kind of age-sensitive MEC models and define the average Age-of-Information (AoI) minimization problems of interests. Then, a novel mixed-policy-based multimodal deep reinforcement learning (RL) framework, called heterogeneous multiagent actor-critic (H-MAAC), is proposed as a paradigm for joint collaboration in the investigated MEC systems, where edge devices and center controller learn the interactive strategies through their own observations. To improve the system performance, we develop the corresponding online algorithm by introducing the edge federated learning mode into the multiagent cooperation whose advantages on learning convergence can be guaranteed theoretically. To the best of our knowledge, it is the first joint MEC collaboration algorithm that combines the edge federated mode with the multiagent actor-critic RL. Furthermore, we evaluate the proposed approach and compare it with popular RL-based methods. As a result, the proposed algorithm not only outperforms the baselines on average system age, but also promotes the stability of training process. Besides, the simulation outcomes provide several insights for collaboration designs over MEC systems.
| Original language | English |
|---|---|
| Pages (from-to) | 1053-1067 |
| Number of pages | 15 |
| Journal | IEEE Internet of Things Journal |
| Volume | 9 |
| Issue number | 2 |
| DOIs | |
| Publication status | Published - 15 Jan 2022 |
Bibliographical note
Publisher Copyright:© 2014 IEEE.
Keywords
- Federated learning (FL)
- joint collaboration
- mixed policies
- mobile-edge computing (MEC)
- multiagent deep reinforcement learning (RL)
- multimodal learning