TY - JOUR
T1 - RTNet
T2 - Relation Transformer Network for Diabetic Retinopathy Multi-Lesion Segmentation
AU - Huang, Shiqi
AU - Li, Jianan
AU - Xiao, Yuze
AU - Shen, Ning
AU - Xu, Tingfa
N1 - Publisher Copyright:
© 1982-2012 IEEE.
PY - 2022/6/1
Y1 - 2022/6/1
N2 - Automatic diabetic retinopathy (DR) lesions segmentation makes great sense of assisting ophthalmologists in diagnosis. Although many researches have been conducted on this task, most prior works paid too much attention to the designs of networks instead of considering the pathological association for lesions. Through investigating the pathogenic causes of DR lesions in advance, we found that certain lesions are closed to specific vessels and present relative patterns to each other. Motivated by the observation, we propose a relation transformer block (RTB) to incorporate attention mechanisms at two main levels: A self-Attention transformer exploits global dependencies among lesion features, while a cross-Attention transformer allows interactions between lesion and vessel features by integrating valuable vascular information to alleviate ambiguity in lesion detection caused by complex fundus structures. In addition, to capture the small lesion patterns first, we propose a global transformer block (GTB) which preserves detailed information in deep network. By integrating the above blocks of dual-branches, our network segments the four kinds of lesions simultaneously. Comprehensive experiments on IDRiD and DDR datasets well demonstrate the superiority of our approach, which achieves competitive performance compared to state-of-The-Arts.
AB - Automatic diabetic retinopathy (DR) lesions segmentation makes great sense of assisting ophthalmologists in diagnosis. Although many researches have been conducted on this task, most prior works paid too much attention to the designs of networks instead of considering the pathological association for lesions. Through investigating the pathogenic causes of DR lesions in advance, we found that certain lesions are closed to specific vessels and present relative patterns to each other. Motivated by the observation, we propose a relation transformer block (RTB) to incorporate attention mechanisms at two main levels: A self-Attention transformer exploits global dependencies among lesion features, while a cross-Attention transformer allows interactions between lesion and vessel features by integrating valuable vascular information to alleviate ambiguity in lesion detection caused by complex fundus structures. In addition, to capture the small lesion patterns first, we propose a global transformer block (GTB) which preserves detailed information in deep network. By integrating the above blocks of dual-branches, our network segments the four kinds of lesions simultaneously. Comprehensive experiments on IDRiD and DDR datasets well demonstrate the superiority of our approach, which achieves competitive performance compared to state-of-The-Arts.
KW - Diabetic retinopathy
KW - deep learning
KW - fundus image
KW - semantic segmentation
KW - transformer
UR - http://www.scopus.com/inward/record.url?scp=85123348337&partnerID=8YFLogxK
U2 - 10.1109/TMI.2022.3143833
DO - 10.1109/TMI.2022.3143833
M3 - Article
C2 - 35041595
AN - SCOPUS:85123348337
SN - 0278-0062
VL - 41
SP - 1596
EP - 1607
JO - IEEE Transactions on Medical Imaging
JF - IEEE Transactions on Medical Imaging
IS - 6
ER -