TY - GEN
T1 - RGBT Decision Stage Fusion Perception Based on Improved YOLOv8
AU - Qin, Yongchun
AU - Ai, Qiang
AU - Zhang, Youtong
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - To enhance the perception capabilities of autonomous vehicles in complex environments, we propose a Visible Light-Infrared (RGBT) decision stage fusion perception method. Based on the YOLOv8 model, a Bidirectional Feature Pyramid Network (BiFPN) structure is introduced, along with the integration of the Convolutional Block Attention Module (CBAM) at three different points for comparative analysis. This optimized model leads to improved detection accuracy. Subsequently, the Non-Maximum Suppression (NMS) algorithm is employed to integrate detection results from both modalities. Experimental results demonstrate that the improved model achieves heightened recognition accuracy on the FLIR dataset for both visible light and infrared scenes, with increases in mAP50 of 1.4% and 2.2% respectively. The enhancement in Fusion perception in mAP50 improved by 8.6% compared to using only visible-light perception., confirming the effectiveness of the visible light-infrared scene perception in challenging environments.
AB - To enhance the perception capabilities of autonomous vehicles in complex environments, we propose a Visible Light-Infrared (RGBT) decision stage fusion perception method. Based on the YOLOv8 model, a Bidirectional Feature Pyramid Network (BiFPN) structure is introduced, along with the integration of the Convolutional Block Attention Module (CBAM) at three different points for comparative analysis. This optimized model leads to improved detection accuracy. Subsequently, the Non-Maximum Suppression (NMS) algorithm is employed to integrate detection results from both modalities. Experimental results demonstrate that the improved model achieves heightened recognition accuracy on the FLIR dataset for both visible light and infrared scenes, with increases in mAP50 of 1.4% and 2.2% respectively. The enhancement in Fusion perception in mAP50 improved by 8.6% compared to using only visible-light perception., confirming the effectiveness of the visible light-infrared scene perception in challenging environments.
KW - autonomous driving perception
KW - decision stage fusion
KW - infrared perception
UR - http://www.scopus.com/inward/record.url?scp=85200680989&partnerID=8YFLogxK
U2 - 10.1109/CVIDL62147.2024.10603614
DO - 10.1109/CVIDL62147.2024.10603614
M3 - Conference contribution
AN - SCOPUS:85200680989
T3 - 2024 5th International Conference on Computer Vision, Image and Deep Learning, CVIDL 2024
SP - 517
EP - 522
BT - 2024 5th International Conference on Computer Vision, Image and Deep Learning, CVIDL 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 5th International Conference on Computer Vision, Image and Deep Learning, CVIDL 2024
Y2 - 19 April 2024 through 21 April 2024
ER -