Delay-aware two-hop cooperative relay communications via approximate MDP and stochastic learning

Research output: Contribution to journalJournal Articlepeer-review

32 Citations (Scopus)

Abstract

In this paper, a low-complexity delay-aware cross-layer scheduling algorithm for two-hop relay communication systems is proposed. The complex interactions of the queues at the source node and the M relay nodes (RSs) are modeled as an infinite horizon average reward Markov decision process (MDP), whose state space involves the joint queue state information (QSI) of the queues at the source node and the M RSs as well as the joint channel state information (CSI) of all S-R and R-D links. To address the curse of dimensionality, an equivalent MDP formulation is first proposed, where the system state depends only on global QSI. Furthermore, using approximate MDP and stochastic learning, an auction-based distributed online learning algorithm is derived, where each node iteratively estimates a per-node value function based on real-time observations of the local CSI and local QSI as well as signaling between relays. The combined distributed learning converges almost surely to a global optimal solution for large arrivals. Finally, it is showed by simulations that the proposed scheme achieves significant gain compared with various baselines such as the conventional CSIT-only control and the throughput optimal control (in stability sense).

Original languageEnglish
Article number6587286
Pages (from-to)7645-7670
Number of pages26
JournalIEEE Transactions on Information Theory
Volume59
Issue number11
DOIs
Publication statusPublished - 2013
Externally publishedYes

Keywords

  • Cooperative communications
  • delay-aware resource allocation
  • distributive algorithm
  • stochastic optimization

Fingerprint

Dive into the research topics of 'Delay-aware two-hop cooperative relay communications via approximate MDP and stochastic learning'. Together they form a unique fingerprint.

Cite this