TY - JOUR
T1 - CTNet
T2 - an efficient coupled transformer network for robust hyperspectral unmixing
AU - Meng, Fanlei
AU - Sun, Haixin
AU - Li, Jie
AU - Xu, Tingfa
N1 - Publisher Copyright:
© 2024 Informa UK Limited, trading as Taylor & Francis Group.
PY - 2024
Y1 - 2024
N2 - This study introduces the coupled transformer Network (CTNet), an architecture designed to enhance the robustness and effectiveness of hyperspectral unmixing (HSU) tasks, addressing key limitations of traditional autoencoder (AE) frameworks. Traditional AEs, consisting of an encoder and a decoder, effectively learn and reconstruct low-dimensional abundance relationships from high-dimensional hyperspectral data but often struggle with spectral variability (SV) and spatial correlations, which can lead to uncertainty in the resulting abundance estimates. CTNet improves upon these limitations by incorporating a two-stream half-Siamese network with an additional encoder trained on pseudo-pure pixels, and further integrates a cross-attention module to leverage global information. This configuration not only guides the AE towards more accurate abundance estimates by directly addressing SV, but also enhances the network’s ability to capture complex spectral information. To minimize the typical reconstruction errors associated with AEs, a transcription loss constraint is applied, which preserves essential details and material-related information often lost during pixel-level reconstruction. Experimental validation on synthetic and three widely-used datasets confirms that CTNet outperforms several state-of-the-art methods, providing a more robust and effective solution for HSU challenges.
AB - This study introduces the coupled transformer Network (CTNet), an architecture designed to enhance the robustness and effectiveness of hyperspectral unmixing (HSU) tasks, addressing key limitations of traditional autoencoder (AE) frameworks. Traditional AEs, consisting of an encoder and a decoder, effectively learn and reconstruct low-dimensional abundance relationships from high-dimensional hyperspectral data but often struggle with spectral variability (SV) and spatial correlations, which can lead to uncertainty in the resulting abundance estimates. CTNet improves upon these limitations by incorporating a two-stream half-Siamese network with an additional encoder trained on pseudo-pure pixels, and further integrates a cross-attention module to leverage global information. This configuration not only guides the AE towards more accurate abundance estimates by directly addressing SV, but also enhances the network’s ability to capture complex spectral information. To minimize the typical reconstruction errors associated with AEs, a transcription loss constraint is applied, which preserves essential details and material-related information often lost during pixel-level reconstruction. Experimental validation on synthetic and three widely-used datasets confirms that CTNet outperforms several state-of-the-art methods, providing a more robust and effective solution for HSU challenges.
KW - Coupled transformer network
KW - cross-attention
KW - robust hyperspectral unmixing
KW - spectral variability
KW - transcription loss
UR - http://www.scopus.com/inward/record.url?scp=85200140527&partnerID=8YFLogxK
U2 - 10.1080/01431161.2024.2371084
DO - 10.1080/01431161.2024.2371084
M3 - Article
AN - SCOPUS:85200140527
SN - 0143-1161
VL - 45
SP - 5679
EP - 5712
JO - International Journal of Remote Sensing
JF - International Journal of Remote Sensing
IS - 17
ER -