TY - GEN
T1 - Infrared and Visible Image Fusion Based on Multiscale Adaptive Transformer
AU - Fei, Erfang
AU - Wang, Yuhao
AU - Zhou, Zhiqiang
AU - Miao, Lingjuan
AU - Li, Jiaqi
AU - Ye, He
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - In our study, we introduce an innovative Transformer-based approach that utilizes multiscale adaptivity for the fusion of infrared and visible images. First of all, we propose a three-branch network structure to extract multiscale differentiated features of source images, and a cross-modal feature interaction module is designed to realize the information interaction of infrared and visible images. And then, inspired by Swin Transformer, a novel adaptive Transformer fusion network is proposed to fuse multiscale features, which fully considers the global information preservation issue during the fusion process and could better integrate the differential and complementary features of infrared and visible images. Furthermore, we present a cross-correlation loss grounded in correlation coefficients to foster a more robust relationship between the fused output and the original images through cross-correlation. The concluding tests reveal that our method's fusion outcomes adeptly harmonize the complementary attributes of various source images, leading to enhanced visual quality and perception.
AB - In our study, we introduce an innovative Transformer-based approach that utilizes multiscale adaptivity for the fusion of infrared and visible images. First of all, we propose a three-branch network structure to extract multiscale differentiated features of source images, and a cross-modal feature interaction module is designed to realize the information interaction of infrared and visible images. And then, inspired by Swin Transformer, a novel adaptive Transformer fusion network is proposed to fuse multiscale features, which fully considers the global information preservation issue during the fusion process and could better integrate the differential and complementary features of infrared and visible images. Furthermore, we present a cross-correlation loss grounded in correlation coefficients to foster a more robust relationship between the fused output and the original images through cross-correlation. The concluding tests reveal that our method's fusion outcomes adeptly harmonize the complementary attributes of various source images, leading to enhanced visual quality and perception.
KW - adaptive Transformer
KW - cross-modal feature interaction
KW - image fusion
KW - multiscale features
UR - http://www.scopus.com/inward/record.url?scp=85189338887&partnerID=8YFLogxK
U2 - 10.1109/CAC59555.2023.10451228
DO - 10.1109/CAC59555.2023.10451228
M3 - Conference contribution
AN - SCOPUS:85189338887
T3 - Proceedings - 2023 China Automation Congress, CAC 2023
SP - 729
EP - 734
BT - Proceedings - 2023 China Automation Congress, CAC 2023
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2023 China Automation Congress, CAC 2023
Y2 - 17 November 2023 through 19 November 2023
ER -