Decision-making models on perceptual uncertainty with distributional reinforcement learning

Shuyuan Xu, Qiao Liu, Yuhui Hu*, Mengtian Xu, Jiachen Hao

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

9 Citations (Scopus)

Abstract

Decision-making for autonomous vehicles in the presence of obstacle occlusions is difficult because the lack of accurate information affects the judgment. Existing methods may lead to overly conservative strategies and time-consuming computations that cannot be balanced with efficiency. We propose to use distributional reinforcement learning to hedge the risk of strategies, optimize the worse cases, and improve the efficiency of the algorithm so that the agent learns better actions. A batch of smaller values is used to replace the average value to optimize the worse case, and combined with frame stacking, we call it Efficient-Fully parameterized Quantile Function (E-FQF). This model is used to evaluate signal-free intersection crossing scenarios and makes more efficient moves and reduces the collision rate compared to conventional reinforcement learning algorithms in the presence of perceived occlusion. The model also has robustness in the case of data loss compared to the method with embedded long and short term memory.

Original languageEnglish
Article number100062
JournalGreen Energy and Intelligent Transportation
Volume2
Issue number2
DOIs
Publication statusPublished - Apr 2023

Keywords

  • Autonomous vehicles
  • Partially observable markov decision process
  • Reinforcement learning
  • Sensing occlusion
  • Unsiganlized intersections

Fingerprint

Dive into the research topics of 'Decision-making models on perceptual uncertainty with distributional reinforcement learning'. Together they form a unique fingerprint.

Cite this