Reinforcement Learning-Based Model Predictive Control for Discrete-Time Systems

Min Lin, Zhongqi Sun*, Yuanqing Xia, Jinhui Zhang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

7 Citations (Scopus)

Abstract

This article proposes a novel reinforcement learning-based model predictive control (RLMPC) scheme for discrete-time systems. The scheme integrates model predictive control (MPC) and reinforcement learning (RL) through policy iteration (PI), where MPC is a policy generator and the RL technique is employed to evaluate the policy. Then the obtained value function is taken as the terminal cost of MPC, thus improving the generated policy. The advantage of doing so is that it rules out the need for the offline design paradigm of the terminal cost, the auxiliary controller, and the terminal constraint in traditional MPC. Moreover, RLMPC proposed in this article enables a more flexible choice of prediction horizon due to the elimination of the terminal constraint, which has great potential in reducing the computational burden. We provide a rigorous analysis of the convergence, feasibility, and stability properties of RLMPC. Simulation results show that RLMPC achieves nearly the same performance as traditional MPC in the control of linear systems and exhibits superiority over traditional MPC for nonlinear ones.

Original languageEnglish
Pages (from-to)3312-3324
Number of pages13
JournalIEEE Transactions on Neural Networks and Learning Systems
Volume35
Issue number3
DOIs
Publication statusPublished - 1 Mar 2024

Keywords

  • Discrete-time systems
  • model predictive control
  • policy iteration (PI)
  • reinforcement learning (RL)

Fingerprint

Dive into the research topics of 'Reinforcement Learning-Based Model Predictive Control for Discrete-Time Systems'. Together they form a unique fingerprint.

Cite this