TY - JOUR
T1 - 3-D Multiobject Tracking With Radar Data Enhancement for Intelligent Vehicles
AU - Li, Luxing
AU - Wei, Chao
AU - Sui, Shuxin
AU - Hu, Jibin
AU - Feng, Fuyong
N1 - Publisher Copyright:
© 2001-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - With the rapid advancement of autonomous driving technology, 3-D multiobject tracking (MOT) has become critical for enhancing the safety of intelligent vehicles. The tracking by detection (TBD) framework improves 3-D tracking performance with simplicity and efficiency. However, existing light detection and ranging (LiDAR)-camera systems struggle in complex scenarios, and millimeter wave radar’s (radar) high-precision motion data remain underutilized. This article presents a multimodal 3-D MOT model that integrates LiDAR, camera, and radar data to overcome these limitations and enhance tracking performance across diverse environments. The proposed model leverages the radar’s high-precision motion data to design an adaptive refinement of confidence scores and adaptive-nonmaximum suppression (A-NMS), improving the accuracy of the preprocessing stage. Additionally, it incorporates LiDAR and camera features to enhance data association and trajectory estimation. A graph attention network (GAT)-based radar modality feature extraction network further improves tracking accuracy by refining motion metrics. Extensive experiments on the nuScenes dataset show that our model achieves a new state-of-the-art (SOTA) tracking accuracy of 74.6% AMOTA, 288 IDS, and 0.514 m AMOTP, along with a real-time performance of 21.8 Hz. Furthermore, in tests of generalizability across various real-vehicle scenarios, our model demonstrates exceptional tracking accuracy, real-time performance, and robustness.
AB - With the rapid advancement of autonomous driving technology, 3-D multiobject tracking (MOT) has become critical for enhancing the safety of intelligent vehicles. The tracking by detection (TBD) framework improves 3-D tracking performance with simplicity and efficiency. However, existing light detection and ranging (LiDAR)-camera systems struggle in complex scenarios, and millimeter wave radar’s (radar) high-precision motion data remain underutilized. This article presents a multimodal 3-D MOT model that integrates LiDAR, camera, and radar data to overcome these limitations and enhance tracking performance across diverse environments. The proposed model leverages the radar’s high-precision motion data to design an adaptive refinement of confidence scores and adaptive-nonmaximum suppression (A-NMS), improving the accuracy of the preprocessing stage. Additionally, it incorporates LiDAR and camera features to enhance data association and trajectory estimation. A graph attention network (GAT)-based radar modality feature extraction network further improves tracking accuracy by refining motion metrics. Extensive experiments on the nuScenes dataset show that our model achieves a new state-of-the-art (SOTA) tracking accuracy of 74.6% AMOTA, 288 IDS, and 0.514 m AMOTP, along with a real-time performance of 21.8 Hz. Furthermore, in tests of generalizability across various real-vehicle scenarios, our model demonstrates exceptional tracking accuracy, real-time performance, and robustness.
KW - 3-D multiobject tracking (MOT)
KW - adaptive enhancement
KW - millimeter wave radar
KW - sensor fusion
UR - http://www.scopus.com/inward/record.url?scp=86000334561&partnerID=8YFLogxK
U2 - 10.1109/JSEN.2025.3545812
DO - 10.1109/JSEN.2025.3545812
M3 - Article
AN - SCOPUS:86000334561
SN - 1530-437X
VL - 25
SP - 16310
EP - 16320
JO - IEEE Sensors Journal
JF - IEEE Sensors Journal
IS - 9
ER -