HV-LIOM: Adaptive Hash-Voxel LiDAR–Inertial SLAM with Multi-Resolution Relocalization and Reinforcement Learning for Autonomous Exploration

  • Shicheng Fan
  • , Xiaopeng Chen*
  • , Weimin Zhang
  • , Peng Xu
  • , Zhengqing Zuo
  • , Xinyan Tan
  • , Xiaohai He
  • , Chandan Sheikder
  • , Meijun Guo
  • , Chengxiang Li
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

This paper presents HV-LIOM (Adaptive Hash-Voxel LiDAR–Inertial Odometry and Mapping), a unified LiDAR–inertial SLAM and autonomous exploration framework for real-time 3D mapping in dynamic, GNSS-denied environments. We propose an adaptive hash-voxel mapping scheme that improves memory efficiency and real-time state estimation by subdividing voxels according to local geometric complexity and point density. To enhance robustness to poor initialization, we introduce a multi-resolution relocalization strategy that enables reliable localization against a prior map under large initial pose errors. A learning-based loop-closure module further detects revisited places and injects global constraints, while global pose-graph optimization maintains long-term map consistency. For autonomous exploration, we integrate a Soft Actor–Critic (SAC) policy that selects informative navigation targets online, improving exploration efficiency in unknown scenes. We evaluate HV-LIOM on public datasets (Hilti and NCLT) and a custom mobile robot platform. Results show that HV-LIOM improves absolute pose accuracy by up to 15.2% over FAST-LIO2 in indoor settings and by 7.6% in large-scale outdoor scenarios. The learned exploration policy achieves comparable or superior area coverage with reduced travel distance and exploration time relative to sampling-based and learning-based baselines.

Original languageEnglish
Article number7558
JournalSensors
Volume25
Issue number24
DOIs
Publication statusPublished - Dec 2025

Keywords

  • Hybrid Voxel Mapping
  • LiDAR-based SLAM
  • active exploration
  • autonomous exploration
  • reinforcement learning

Fingerprint

Dive into the research topics of 'HV-LIOM: Adaptive Hash-Voxel LiDAR–Inertial SLAM with Multi-Resolution Relocalization and Reinforcement Learning for Autonomous Exploration'. Together they form a unique fingerprint.

Cite this