TY - GEN
T1 - Application Barrier-Free Deep Reinforcement Learning-Based Navigation in Mapless Environment
AU - Yu, Jin
AU - Wang, Zhengjie
AU - Wu, Yanxuan
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
PY - 2022
Y1 - 2022
N2 - This paper presents a case study of a novel model-free training framework for indoor mapless navigation without any prior expert demonstrations. A model-free Deep Reinforcement Learning (DRL) approach is implemented, which is capable of good sample efficiency and convergence rate. The navigation policy learns directly end-to-end from raw lidar information in a virtual environment and can be deployed directly on a real mobile robot platform. Without any map or self-location information, the robot is only provided with raw lidar information and the relative position to the target calculated by the inertial measurement unit and encoder which can achieve a large number of success of collision-free navigation tasks during the real-world test. We present a thorough evaluation of different input information, different reward functions, different DRL algorithm structures and different training environments both in simulation and real world. The results show that with appropriate reward function, DRL algorithm structures and environments, the trained agent is able to reach the target without collision. In addition, the policy deployed from virtual to the real robot performs better after processing the input information..
AB - This paper presents a case study of a novel model-free training framework for indoor mapless navigation without any prior expert demonstrations. A model-free Deep Reinforcement Learning (DRL) approach is implemented, which is capable of good sample efficiency and convergence rate. The navigation policy learns directly end-to-end from raw lidar information in a virtual environment and can be deployed directly on a real mobile robot platform. Without any map or self-location information, the robot is only provided with raw lidar information and the relative position to the target calculated by the inertial measurement unit and encoder which can achieve a large number of success of collision-free navigation tasks during the real-world test. We present a thorough evaluation of different input information, different reward functions, different DRL algorithm structures and different training environments both in simulation and real world. The results show that with appropriate reward function, DRL algorithm structures and environments, the trained agent is able to reach the target without collision. In addition, the policy deployed from virtual to the real robot performs better after processing the input information..
KW - Deep reinforcement learning
KW - End-to-end planning
KW - Mapless navigation
KW - Real-platform deployment
UR - http://www.scopus.com/inward/record.url?scp=85118098361&partnerID=8YFLogxK
U2 - 10.1007/978-981-16-6324-6_65
DO - 10.1007/978-981-16-6324-6_65
M3 - Conference contribution
AN - SCOPUS:85118098361
SN - 9789811663239
T3 - Lecture Notes in Electrical Engineering
SP - 643
EP - 653
BT - Proceedings of 2021 Chinese Intelligent Systems Conference - Volume II
A2 - Jia, Yingmin
A2 - Zhang, Weicun
A2 - Fu, Yongling
A2 - Yu, Zhiyuan
A2 - Zheng, Song
PB - Springer Science and Business Media Deutschland GmbH
T2 - 17th Chinese Intelligent Systems Conference, CISC 2021
Y2 - 16 October 2021 through 17 October 2021
ER -