Autonomous Vehicle Dynamic Following Based on Improved Reinforcement Learning

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Target following technology, as one of the core technologies for autonomous vehicles, enables vehicles to automatically track dynamic targets. This paper proposes a path planning algorithm that combines proximal policy optimization(PPO) with Model predictive control(MPC) for dynamic target following tasks of autonomous vehicles in unknown environments. The algorithm utilizes neural networks to reduce computational load and employs MPC to compensate for the shortcomings of PPO in the prediction phase, thereby enhancing the following and obstacle avoidance capabilities of autonomous vehicles in dynamic environments. Simulation experiments show that the proposed PPO-MPC improves training speed by 35%, achieves trajectory following effects in low-speed dynamic target following tasks, and reduces the time required for static and dynamic obstacle avoidance tasks by approximately 8.5% compared to the original PPO, with smoother trajectories. It is capable of performing dynamic target following tasks in unknown environments with multiple obstacles.

Original languageEnglish
Title of host publicationProceedings of the 44th Chinese Control Conference, CCC 2025
EditorsJian Sun, Hongpeng Yin
PublisherIEEE Computer Society
Pages4323-4328
Number of pages6
ISBN (Electronic)9789887581611
DOIs
Publication statusPublished - 2025
Externally publishedYes
Event44th Chinese Control Conference, CCC 2025 - Chongqing, China
Duration: 28 Jul 202530 Jul 2025

Publication series

NameChinese Control Conference, CCC
ISSN (Print)1934-1768
ISSN (Electronic)2161-2927

Conference

Conference44th Chinese Control Conference, CCC 2025
Country/TerritoryChina
CityChongqing
Period28/07/2530/07/25

Keywords

  • Autonomous vehicles
  • Dynamic target following
  • Model predictive control
  • Proximal Policy Optimization

Fingerprint

Dive into the research topics of 'Autonomous Vehicle Dynamic Following Based on Improved Reinforcement Learning'. Together they form a unique fingerprint.

Cite this