GCANet: A Cross-Modal Pedestrian Detection Method Based on Gaussian Cross Attention Network

Peiran Peng, Feng Mu, Peilin Yan, Liqiang Song, Hui Li, Yu Chen, Jianan Li, Tingfa Xu*

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Citation (Scopus)

Abstract

Pedestrian detection is a critical but challenging research field widely applicable in self-driving, surveillance and robotics. The performance of pedestrian detection is not ideal under the limitation of imaging conditions, especially at night or occlusion. To overcome these obstacles, we propose a cross-modal pedestrian detection network based on Gaussian Cross Attention (GCANet) improving the detection performance by a full use of multi-modal features. Through the bidirectional coupling of local features of different modals, the feature interaction and fusion between different modals are realized, and the salient features between multi-modal are effectively emphasized, thus improving the detection accuracy. Experimental results demonstrate GCANet achieves the highest accuracy with the state-of-the-art on KAIST multi-modal pedestrian dataset.

Original languageEnglish
Title of host publicationIntelligent Computing - Proceedings of the 2022 Computing Conference
EditorsKohei Arai
PublisherSpringer Science and Business Media Deutschland GmbH
Pages520-530
Number of pages11
ISBN (Print)9783031104633
DOIs
Publication statusPublished - 2022
EventComputing Conference, 2022 - Virtual, Online
Duration: 14 Jul 202215 Jul 2022

Publication series

NameLecture Notes in Networks and Systems
Volume507 LNNS
ISSN (Print)2367-3370
ISSN (Electronic)2367-3389

Conference

ConferenceComputing Conference, 2022
CityVirtual, Online
Period14/07/2215/07/22

Keywords

  • Gaussian Cross Attention
  • Multi-modal fusion
  • Pedestrian detection

Fingerprint

Dive into the research topics of 'GCANet: A Cross-Modal Pedestrian Detection Method Based on Gaussian Cross Attention Network'. Together they form a unique fingerprint.

Cite this