Abstract
Reinforcement Learning Model Predictive Control (RL-MPC) has achieved significant progress in recent years. However, existing approaches still have some limitations. This paper proposes a Bayesian policy exploration method for RLMPC that substantially enhances its performance. Specifically, we implement Bayesian posterior estimation of value functions and introduce an optimistic exploration strategy tailored for efficient exploration of RL-MPC, which improves the sample efficiency of RL policy exploration. Then an optimistic Bayesian exploration strategy is proposed, which encourages the agent to leverage existing model information to achieve superior control performance. The soundness and effectiveness of our method are evaluated through an empirical study of controlling a drone to reach targets subject to uncertain model parameters and environmental perturbations. The results validate that our approach has superior performance compared with benchmarks.
| Original language | English |
|---|---|
| Title of host publication | 2025 IEEE 19th International Conference on Control and Automation, ICCA 2025 |
| Publisher | IEEE Computer Society |
| Pages | 460-465 |
| Number of pages | 6 |
| ISBN (Electronic) | 9798331595593 |
| ISBN (Print) | 9798331595609 |
| DOIs | |
| Publication status | Published - 2 Sept 2025 |
| Externally published | Yes |
| Event | 19th IEEE International Conference on Control and Automation, ICCA 2025 - Tallinn, Estonia Duration: 30 Jun 2025 → 3 Jul 2025 |
Publication series
| Name | IEEE International Conference on Control and Automation, ICCA |
|---|---|
| ISSN (Print) | 1948-3449 |
| ISSN (Electronic) | 1948-3457 |
Conference
| Conference | 19th IEEE International Conference on Control and Automation, ICCA 2025 |
|---|---|
| Country/Territory | Estonia |
| City | Tallinn |
| Period | 30/06/25 → 3/07/25 |
Bibliographical note
Publisher Copyright:© 2025 IEEE.
Keywords
- reinforcement learning
- model predictive control
- Bayesian estimation
- optimistic policy exploration