An Obstacle Avoidance Method Using Asynchronous Policy-based Deep Reinforcement Learning with Discrete Action

Yuechuan Wang*, Fenxi Yao*, Lingguo Cui, Senchun Chai

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

4 Citations (Scopus)

Abstract

With the increasing application of mobile robots in manufacturing, service, and military fields, the demand of intelligent autonomous decision is also growing. In this paper, the state-of-the-art policy-based deep reinforcement learning (DRL) algorithm is applied to the mobile robot obstacle avoidance task. To solve the strong coupling in the existing DRL-based training process of mobile robot decision, an asynchronous decoupling architecture is proposed in this paper, which greatly improves the scalability and sample generation efficiency of the DRL algorithm. At the same time, based on the asynchronous decoupling architecture and Soft Actor-Critic (SAC) algorithm, we design an obstacle avoidance method named asynchronous dueling-based discrete action SAC (ADDSAC). Experiments show both action discretization and dueling network are simple and powerful techniques to improve obstacle avoidance performance. Finally, the reasons are also analyzed from different perspectives.

Original languageEnglish
Title of host publicationProceedings of the 34th Chinese Control and Decision Conference, CCDC 2022
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages6235-6241
Number of pages7
ISBN (Electronic)9781665478960
DOIs
Publication statusPublished - 2022
Event34th Chinese Control and Decision Conference, CCDC 2022 - Hefei, China
Duration: 15 Aug 202217 Aug 2022

Publication series

NameProceedings of the 34th Chinese Control and Decision Conference, CCDC 2022

Conference

Conference34th Chinese Control and Decision Conference, CCDC 2022
Country/TerritoryChina
CityHefei
Period15/08/2217/08/22

Keywords

  • Mobile robot
  • Soft Actor-Critic
  • deep reinforcement learning
  • obstacle avoidance

Fingerprint

Dive into the research topics of 'An Obstacle Avoidance Method Using Asynchronous Policy-based Deep Reinforcement Learning with Discrete Action'. Together they form a unique fingerprint.

Cite this