Hybrid bidirectional rapidly exploring random tree path planning algorithm with reinforcement learning

Junkui Wang, Kaoru Hirota, Xiangdong Wu, Yaping Dai, Zhiyang Jia*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

10 Citations (Scopus)

Abstract

The randomness of path generation and slow convergence to the optimal path are two major problems in the current rapidly exploring random tree (RRT) path planning algorithm. Herein, a novel reinforcement-learning-based hybrid bidirectional rapidly exploring random tree (H-BRRT) is presented to solve these problems. To model the random exploration process, a target gravitational strategy is introduced. Reinforcement learning is applied to the improved target gravitational strategy using two operations: random exploration and target gravitational exploration. The algorithm is controlled to switch operations adaptively according to the accumulated performance. It not only improves the search efficiency, but also shortens the generated path after the proposed strategy is applied to a bidirectional rapidly exploring random tree (BRRT). In addition, to solve the problem of the traditional RRT continuously falling into the local optimum, an improved exploration strategy with collision weight is applied to the BRRT. Experimental results implemented in a robot operating system indicate that the proposed H-BRRT significantly outperforms alternative approaches such as the RRT and BRRT. The proposed algorithm enhances the capability of identifying unknown spaces and avoiding local optima.

Original languageEnglish
Pages (from-to)121-129
Number of pages9
JournalJournal of Advanced Computational Intelligence and Intelligent Informatics
Volume25
Issue number1
DOIs
Publication statusPublished - 20 Jan 2021

Keywords

  • Path planning
  • Q-learning
  • Rapidly exploring random tree
  • Reinforcement learning
  • Target gravitational strategy

Fingerprint

Dive into the research topics of 'Hybrid bidirectional rapidly exploring random tree path planning algorithm with reinforcement learning'. Together they form a unique fingerprint.

Cite this