TY - JOUR
T1 - CSTSUNet
T2 - A Cross Swin Transformer-Based Siamese U-Shape Network for Change Detection in Remote Sensing Images
AU - Wu, Yaping
AU - Li, Lu
AU - Wang, Nan
AU - Li, Wei
AU - Fan, Junfang
AU - Tao, Ran
AU - Wen, Xin
AU - Wang, Yanfeng
N1 - Publisher Copyright:
© 1980-2012 IEEE.
PY - 2023
Y1 - 2023
N2 - Change detection (CD) in remote sensing (RS) images is a critical task that has achieved significant success by deep learning. Current networks often employ pixel-based differencing, proportion, classification-based, or feature concatenation methods to represent changes of interest. However, these methods fail to effectively detect the desired changes, as they are highly sensitive to factors such as atmospheric conditions, lighting variations, and phenological variations, resulting in detection errors. Inspired by the transformer structure, we adopt a cross-attention mechanism to more robustly extract feature differences between bitemporal images. The motivation of the method is based on the assumption that if there is no change between image pairs, the semantic features from one temporal image can well be represented by the semantic features from another temporal image. Conversely if there is a change, there are significant reconstruction errors. Therefore, a Cross Swin transformer-based Siamese U-shaped network namely CSTSUNet is proposed for RS CD. CSTSUnet consists of encoder, difference feature extraction, and decoder. The encoder is based on a hierarchical residual network (ResNet) with the Siamese U-net structure, allowing parallel processing of bitemporal images and extraction of multiscale features. The difference feature extraction consists of four difference feature extraction modules that compute difference feature at multiple scales. In this module, Cross Swin transformer is employed in each difference feature extraction module to communicate the information of bitemporal images. The decoder takes in the multiscale difference features as input, injects details and boundaries iteratively level by level, and makes the change map more and more accurate. We conduct experiments on three public datasets, and the experimental results demonstrate that the proposed CSTSUNet outperforms other state-of-the-art methods in terms of both qualitative and quantitative analyses. Our code is available at https://github.com/l7170/CSTSUNet.git.
AB - Change detection (CD) in remote sensing (RS) images is a critical task that has achieved significant success by deep learning. Current networks often employ pixel-based differencing, proportion, classification-based, or feature concatenation methods to represent changes of interest. However, these methods fail to effectively detect the desired changes, as they are highly sensitive to factors such as atmospheric conditions, lighting variations, and phenological variations, resulting in detection errors. Inspired by the transformer structure, we adopt a cross-attention mechanism to more robustly extract feature differences between bitemporal images. The motivation of the method is based on the assumption that if there is no change between image pairs, the semantic features from one temporal image can well be represented by the semantic features from another temporal image. Conversely if there is a change, there are significant reconstruction errors. Therefore, a Cross Swin transformer-based Siamese U-shaped network namely CSTSUNet is proposed for RS CD. CSTSUnet consists of encoder, difference feature extraction, and decoder. The encoder is based on a hierarchical residual network (ResNet) with the Siamese U-net structure, allowing parallel processing of bitemporal images and extraction of multiscale features. The difference feature extraction consists of four difference feature extraction modules that compute difference feature at multiple scales. In this module, Cross Swin transformer is employed in each difference feature extraction module to communicate the information of bitemporal images. The decoder takes in the multiscale difference features as input, injects details and boundaries iteratively level by level, and makes the change map more and more accurate. We conduct experiments on three public datasets, and the experimental results demonstrate that the proposed CSTSUNet outperforms other state-of-the-art methods in terms of both qualitative and quantitative analyses. Our code is available at https://github.com/l7170/CSTSUNet.git.
KW - Change detection (CD)
KW - deep learning
KW - remote sensing (RS) image
KW - transformer
UR - http://www.scopus.com/inward/record.url?scp=85176300304&partnerID=8YFLogxK
U2 - 10.1109/TGRS.2023.3326813
DO - 10.1109/TGRS.2023.3326813
M3 - Article
AN - SCOPUS:85176300304
SN - 0196-2892
VL - 61
JO - IEEE Transactions on Geoscience and Remote Sensing
JF - IEEE Transactions on Geoscience and Remote Sensing
M1 - 5623715
ER -