TY - GEN
T1 - Infrared Target Detection Using Intensity Saliency and Self-Attention
AU - Zhang, Ruiheng
AU - Xu, Min
AU - Shi, Yaxin
AU - Fan, Jian
AU - Mu, Chengpo
AU - Xu, Lixin
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/10
Y1 - 2020/10
N2 - Infrared target detection is essential for many computer vision tasks. Generally, the IR images present common infrared characteristics, such as poor texture information, low resolution, and high noise. However, these characteristics are ignored in the existing detection methods, making them fail in real-world scenarios. In this paper, we take infrared intensity into account and propose a novel backbone network named Deep-IRTarget. We first extract infrared intensity saliency by a convolution with a Gaussian kernel filtering the images in the frequency domain. We then propose the triple self-attention network to further extract spatial domain image saliency by selectively emphasize interdependent semantic features in each channel. Jointly exploiting infrared characteristics in the frequency domain and the overall semantic interdependencies in the spatial domain, the proposed Deep-IRTarget outperforms existing methods in real-world Infrared target detection tasks. Experimental results on two infrared imagery datasets demonstrate the superiorly of our model.
AB - Infrared target detection is essential for many computer vision tasks. Generally, the IR images present common infrared characteristics, such as poor texture information, low resolution, and high noise. However, these characteristics are ignored in the existing detection methods, making them fail in real-world scenarios. In this paper, we take infrared intensity into account and propose a novel backbone network named Deep-IRTarget. We first extract infrared intensity saliency by a convolution with a Gaussian kernel filtering the images in the frequency domain. We then propose the triple self-attention network to further extract spatial domain image saliency by selectively emphasize interdependent semantic features in each channel. Jointly exploiting infrared characteristics in the frequency domain and the overall semantic interdependencies in the spatial domain, the proposed Deep-IRTarget outperforms existing methods in real-world Infrared target detection tasks. Experimental results on two infrared imagery datasets demonstrate the superiorly of our model.
KW - Attention
KW - CNN
KW - Fourier transform
KW - Infrared target detection
UR - http://www.scopus.com/inward/record.url?scp=85098664707&partnerID=8YFLogxK
U2 - 10.1109/ICIP40778.2020.9191055
DO - 10.1109/ICIP40778.2020.9191055
M3 - Conference contribution
AN - SCOPUS:85098664707
T3 - Proceedings - International Conference on Image Processing, ICIP
SP - 1991
EP - 1995
BT - 2020 IEEE International Conference on Image Processing, ICIP 2020 - Proceedings
PB - IEEE Computer Society
T2 - 2020 IEEE International Conference on Image Processing, ICIP 2020
Y2 - 25 September 2020 through 28 September 2020
ER -