TY - GEN
T1 - Real-Data-Driven Offline Reinforcement Learning for Autonomous Vehicle Speed Decision Making
AU - Hao, Jiachen
AU - Xu, Shuyuan
AU - Chen, Xuemei
AU - Fu, Shuaiqi
AU - Yang, Dongqing
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - In recent years, reinforcement learning has demonstrated its powerful learning ability and application potential in the autonomous driving decision module, in comparison to traditional methods. However, due to the interaction with the environment, it is mostly limited to the simulation environment. Offline RL, on the other hand, can learn through fixed data sets and is considered a feasible means to push reinforcement learning to practical applications. To verify whether offline RL can exhibit excellent performance under real-world datasets, we provide a benchmark for offline RL using the Argo dataset. This paper first introduces a variety of offline RL algorithms, followed by the processing of scenes such as starting, following, and parking in the dataset, as well as the generation of the dataset and simulation environment. Different datasets were obtained through data enhancement and other methods, and several state-of-the-art(SOTA) offline RL algorithms were tested and compared with imitation learning BC. Finally, the conclusion and analysis are presented. Our code is available at https://github.com/hjcwuhuqifei/offline-rl-benchmark-by-argo.
AB - In recent years, reinforcement learning has demonstrated its powerful learning ability and application potential in the autonomous driving decision module, in comparison to traditional methods. However, due to the interaction with the environment, it is mostly limited to the simulation environment. Offline RL, on the other hand, can learn through fixed data sets and is considered a feasible means to push reinforcement learning to practical applications. To verify whether offline RL can exhibit excellent performance under real-world datasets, we provide a benchmark for offline RL using the Argo dataset. This paper first introduces a variety of offline RL algorithms, followed by the processing of scenes such as starting, following, and parking in the dataset, as well as the generation of the dataset and simulation environment. Different datasets were obtained through data enhancement and other methods, and several state-of-the-art(SOTA) offline RL algorithms were tested and compared with imitation learning BC. Finally, the conclusion and analysis are presented. Our code is available at https://github.com/hjcwuhuqifei/offline-rl-benchmark-by-argo.
KW - autonomous driving
KW - offline rl
KW - speed decision
UR - http://www.scopus.com/inward/record.url?scp=85200399338&partnerID=8YFLogxK
U2 - 10.1109/CCDC62350.2024.10588065
DO - 10.1109/CCDC62350.2024.10588065
M3 - Conference contribution
AN - SCOPUS:85200399338
T3 - Proceedings of the 36th Chinese Control and Decision Conference, CCDC 2024
SP - 2504
EP - 2511
BT - Proceedings of the 36th Chinese Control and Decision Conference, CCDC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 36th Chinese Control and Decision Conference, CCDC 2024
Y2 - 25 May 2024 through 27 May 2024
ER -