From perturbation analysis to Markov decision processes and reinforcement learning

Xi Ren Cao*

*Corresponding author for this work

Research output: Contribution to journalJournal Articlepeer-review

49 Citations (Scopus)

Abstract

The goals of perturbation analysis (PA), Markov decision processes (MDPs), and reinforcement learning (RL) are common: to make decisions to improve the system performance based on the information obtained by analyzing the current system behavior. In this paper, we study the relations among these closely related fields. We show that MDP solutions can be derived naturally from performance sensitivity analysis provided by PA. Performance potential plays an important role in both PA and MDPs; it also offers a clear intuitive interpretation for many results. Reinforcement learning, TD(λ), neuro-dynamic programming, etc., are efficient ways of estimating the performance potentials and related quantities based on sample paths. The sensitivity point of view of PA, MDP. and RL brings in some new insight to the area of learning and optimization. In particular, gradient-based optimization can be applied to parameterized systems with large state spaces, and gradient-based policy iteration can be applied to some nonstandard MDPs such as systems with correlated actions, etc. Potential-based on-line approaches and their advantages are also discussed.

Original languageEnglish
Pages (from-to)9-39
Number of pages31
JournalDiscrete Event Dynamic Systems: Theory and Applications
Volume13
Issue number1-2
DOIs
Publication statusPublished - Jan 2003

Keywords

  • Gradient-based policy iteration
  • Perturbation realization
  • Poisson equations
  • Potentials
  • Q-learning
  • TD(λ)

Fingerprint

Dive into the research topics of 'From perturbation analysis to Markov decision processes and reinforcement learning'. Together they form a unique fingerprint.

Cite this