Cross-modal attention network for retinal disease classification based on multi-modal images

Zirong Liu, Yan Hu, Zhongxi Qiu, Yanyan Niu, Dan Zhou, Xiaoling Li, Junyong Shen, Hongyang Jiang, Heng Li, Jiang Liu*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Multi-modal eye disease screening improves diagnostic accuracy by providing lesion information from different sources. However, existing multi-modal automatic diagnosis methods tend to focus on the specificity of modalities and ignore the spatial correlation of images. This paper proposes a novel cross-modal retinal disease diagnosis network (CRD-Net) that digs out the relevant features from modal images aided for multiple retinal disease diagnosis. Specifically, our model introduces a cross-modal attention (CMA) module to query and adaptively pay attention to the relevant features of the lesion in the different modal images. In addition, we also propose multiple loss functions to fuse features with modality correlation and train a multi-modal retinal image classification network to achieve a more accurate diagnosis. Experimental evaluation on three publicly available datasets shows that our CRD-Net outperforms existing single-modal and multi-modal methods, demonstrating its superior performance.

Original languageEnglish
Pages (from-to)3699-3714
Number of pages16
JournalBiomedical Optics Express
Volume15
Issue number6
DOIs
Publication statusPublished - 1 Jun 2024
Externally publishedYes

Fingerprint

Dive into the research topics of 'Cross-modal attention network for retinal disease classification based on multi-modal images'. Together they form a unique fingerprint.

Cite this