TY - GEN
T1 - GDAFormer
T2 - 21st International Conference on Intelligent Computing, ICIC 2025
AU - Fan, Haoran
AU - Yu, Xiangyang
AU - Li, Heng
AU - Li, Haojin
AU - Liu, Jiang
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - Fundus image enhancement plays a crucial role in retinal disease diagnosis but remains hindered by challenges such as uneven illumination, low contrast, and domain variability. We present GDAFormer, a novel transformer-based framework tailored for fundus image enhancement, incorporating multi-head depth-wise convolutional self-attention and a gated dual-attention fusion block. This design captures long-range dependencies while adaptively integrating global structures and local pathological features, preserving vascular continuity and mitigating over-smoothing. A segmentation-aware loss, guided by a pretrained retinal segmentation network, further reinforces structural fidelity without requiring paired supervision. Extensive evaluations on the FIQ, RCF, and RF datasets show that GDAFormer consistently outperforms state-of-the-art methods in both PSNR and SSIM metrics. Our approach achieves strong generalization across diverse imaging protocols, making it a robust and clinically meaningful enhancement tool.
AB - Fundus image enhancement plays a crucial role in retinal disease diagnosis but remains hindered by challenges such as uneven illumination, low contrast, and domain variability. We present GDAFormer, a novel transformer-based framework tailored for fundus image enhancement, incorporating multi-head depth-wise convolutional self-attention and a gated dual-attention fusion block. This design captures long-range dependencies while adaptively integrating global structures and local pathological features, preserving vascular continuity and mitigating over-smoothing. A segmentation-aware loss, guided by a pretrained retinal segmentation network, further reinforces structural fidelity without requiring paired supervision. Extensive evaluations on the FIQ, RCF, and RF datasets show that GDAFormer consistently outperforms state-of-the-art methods in both PSNR and SSIM metrics. Our approach achieves strong generalization across diverse imaging protocols, making it a robust and clinically meaningful enhancement tool.
KW - Fundus Image Enhancement
KW - Structure Preservation
KW - Vision Transformer
UR - https://www.scopus.com/pages/publications/105012357686
U2 - 10.1007/978-981-96-9863-9_20
DO - 10.1007/978-981-96-9863-9_20
M3 - Conference contribution
AN - SCOPUS:105012357686
SN - 9789819698622
T3 - Lecture Notes in Computer Science
SP - 234
EP - 244
BT - Advanced Intelligent Computing Technology and Applications - 21st International Conference, ICIC 2025, Proceedings
A2 - Huang, De-Shuang
A2 - Pan, Yijie
A2 - Chen, Wei
A2 - Li, Bo
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 26 July 2025 through 29 July 2025
ER -