TY - GEN
T1 - An Obstacle Avoidance Method Using Asynchronous Policy-based Deep Reinforcement Learning with Discrete Action
AU - Wang, Yuechuan
AU - Yao, Fenxi
AU - Cui, Lingguo
AU - Chai, Senchun
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - With the increasing application of mobile robots in manufacturing, service, and military fields, the demand of intelligent autonomous decision is also growing. In this paper, the state-of-the-art policy-based deep reinforcement learning (DRL) algorithm is applied to the mobile robot obstacle avoidance task. To solve the strong coupling in the existing DRL-based training process of mobile robot decision, an asynchronous decoupling architecture is proposed in this paper, which greatly improves the scalability and sample generation efficiency of the DRL algorithm. At the same time, based on the asynchronous decoupling architecture and Soft Actor-Critic (SAC) algorithm, we design an obstacle avoidance method named asynchronous dueling-based discrete action SAC (ADDSAC). Experiments show both action discretization and dueling network are simple and powerful techniques to improve obstacle avoidance performance. Finally, the reasons are also analyzed from different perspectives.
AB - With the increasing application of mobile robots in manufacturing, service, and military fields, the demand of intelligent autonomous decision is also growing. In this paper, the state-of-the-art policy-based deep reinforcement learning (DRL) algorithm is applied to the mobile robot obstacle avoidance task. To solve the strong coupling in the existing DRL-based training process of mobile robot decision, an asynchronous decoupling architecture is proposed in this paper, which greatly improves the scalability and sample generation efficiency of the DRL algorithm. At the same time, based on the asynchronous decoupling architecture and Soft Actor-Critic (SAC) algorithm, we design an obstacle avoidance method named asynchronous dueling-based discrete action SAC (ADDSAC). Experiments show both action discretization and dueling network are simple and powerful techniques to improve obstacle avoidance performance. Finally, the reasons are also analyzed from different perspectives.
KW - Mobile robot
KW - Soft Actor-Critic
KW - deep reinforcement learning
KW - obstacle avoidance
UR - http://www.scopus.com/inward/record.url?scp=85149552149&partnerID=8YFLogxK
U2 - 10.1109/CCDC55256.2022.10033562
DO - 10.1109/CCDC55256.2022.10033562
M3 - Conference contribution
AN - SCOPUS:85149552149
T3 - Proceedings of the 34th Chinese Control and Decision Conference, CCDC 2022
SP - 6235
EP - 6241
BT - Proceedings of the 34th Chinese Control and Decision Conference, CCDC 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 34th Chinese Control and Decision Conference, CCDC 2022
Y2 - 15 August 2022 through 17 August 2022
ER -