摘要
Camera and lidar are the key sources of information in autonomous vehicles (AVs). However, in the current 3D object detection tasks, most of the pure point cloud network detection capabilities are better than those of image and laser point cloud fusion networks. Existing studies summarize the reasons for this as the misalignment of view between image and radar information and the difficulty of matching heterogeneous features. Single-stage fusion algorithm is difficult to fully fuse the features of both. For this reason, a nova 3D object detection based on multilayer multimodal fusion (3DMMF) is presented. First, in the early-fusion phase, point clouds are encoded locally by Frustum-RGB-PointPainting (FRP) formed by the 2D detection frame. Then, the encoded point cloud input is combined with the self-attention mechanism context-aware channel to expand the PointPillars detection network. In the later-fusion phase, 2D and 3D candidate boxes are coded as two sets of sparse tensors before they are not greatly suppressed, and the final 3D target detection result is obtained by using the camera lidar object candidates fusion (CLOCs) network. Experiments on KITTI datasets show that this fusion detection method has a significant performance improvement over the baseline of pure point cloud networks, with an average mAP improvement of 6.24%.
投稿的翻译标题 | 3D Object Detection Based on Multilayer Multimodal Fusion |
---|---|
源语言 | 繁体中文 |
页(从-至) | 696-708 |
页数 | 13 |
期刊 | Tien Tzu Hsueh Pao/Acta Electronica Sinica |
卷 | 52 |
期 | 3 |
DOI | |
出版状态 | 已出版 - 3月 2024 |
关键词
- 3D target detection
- auto-driving
- multi-sensor fusion
- point cloud coding
- self-attention mechanism