TY - JOUR
T1 - GRC-Net
T2 - Fusing GAT-Based 4D Radar and Camera for 3D Object Detection
AU - Fan, Lili
AU - Zeng, Changxian
AU - Li, Yunjie
AU - Wang, Xu
AU - Cao, Dongpu
N1 - Publisher Copyright:
© 2023 SAE International.
PY - 2023/12/31
Y1 - 2023/12/31
N2 - The fusion of multi-modal perception in autonomous driving plays a pivotal role in vehicle behavior decision-making. However, much of the previous research has predominantly focused on the fusion of Lidar and cameras. Although Lidar offers an ample supply of point cloud data, its high cost and the substantial volume of point cloud data can lead to computational delays. Consequently, investigating perception fusion under the context of 4D millimeter-wave radar is of paramount importance for cost reduction and enhanced safety. Nevertheless, 4D millimeter-wave radar faces challenges including sparse point clouds, limited information content, and a lack of fusion strategies. In this paper, we introduce, for the first time, an approach that leverages Graph Neural Networks to assist in expressing features from 4D millimeter-wave radar point clouds. This approach effectively extracts unstructured point cloud features, addressing the loss of object detection due to sparsity. Additionally, we propose the Multi-Modal Fusion Module (MMFM), which aligns and fuses features from graphs, radar pseudo-images generated from Pillars, and camera images within a geometric space. We validate our model using the View-of-Delft (VoD) dataset. Experimental results demonstrate that the proposed method efficiently fuses camera and 4D radar features, resulting in enhanced 3D detection performance.
AB - The fusion of multi-modal perception in autonomous driving plays a pivotal role in vehicle behavior decision-making. However, much of the previous research has predominantly focused on the fusion of Lidar and cameras. Although Lidar offers an ample supply of point cloud data, its high cost and the substantial volume of point cloud data can lead to computational delays. Consequently, investigating perception fusion under the context of 4D millimeter-wave radar is of paramount importance for cost reduction and enhanced safety. Nevertheless, 4D millimeter-wave radar faces challenges including sparse point clouds, limited information content, and a lack of fusion strategies. In this paper, we introduce, for the first time, an approach that leverages Graph Neural Networks to assist in expressing features from 4D millimeter-wave radar point clouds. This approach effectively extracts unstructured point cloud features, addressing the loss of object detection due to sparsity. Additionally, we propose the Multi-Modal Fusion Module (MMFM), which aligns and fuses features from graphs, radar pseudo-images generated from Pillars, and camera images within a geometric space. We validate our model using the View-of-Delft (VoD) dataset. Experimental results demonstrate that the proposed method efficiently fuses camera and 4D radar features, resulting in enhanced 3D detection performance.
KW - 3D object detection
KW - 4D radar
KW - Autonomous driving
KW - multimodal fusion
UR - https://www.scopus.com/pages/publications/105011631388
U2 - 10.4271/2023-01-7088
DO - 10.4271/2023-01-7088
M3 - Article
AN - SCOPUS:105011631388
SN - 2641-9645
VL - 6
SP - 2690
EP - 2696
JO - SAE International Journal of Advances and Current Practices in Mobility
JF - SAE International Journal of Advances and Current Practices in Mobility
IS - 5
ER -