TY - JOUR
T1 - DMDMN
T2 - distraction mining and dual mutual network for curvilinear structure segmentation
AU - Wu, Qingbo
AU - Meng, Shuofei
AU - Huang, Wei
AU - Chen, Shengyong
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2025.
PY - 2025/7
Y1 - 2025/7
N2 - Curvilinear Structure Segmentation (CSS) aims to predict the binary masks of curvilinear objects in photographs, such as blood vessels and road cracks. Despite impressive segmentation results, current CSS methods often struggle with disconnected segmentations and inaccurate edges, essentially caused by the distractions in the feature domain, which are from image domain-specific challenges, and by the imbalance of scales of thicker and thinner branches. To address these problems, we propose Distraction Mining and Dual Mutual Network (DMDMN) based on reverse attention and mutual learning to improve CSS. Specifically, DMDMN first extracts various levels of features from an input image using a plain network. Then, a Reverse Attention Module (RAM) is designed at each level to enhance the extracted features by identifying and removing false positive and negative distractions. Next is a Three-Head Fusion Module (THFM) at a separate level, which serves as an exchanger to mutually integrate the features from the branch segmentation head with those from the heads of skeleton extraction and edge detection. With the segmented results flowing back to RAMs, the previous two steps alternate several times till the final prediction of a high-quality segmentation, which is topologically connective along branches and pixel-wise accurate at edges. In addition, extensive experiments on five public datasets have demonstrated the superiority of the proposed DMDMN over state-of-the-art approaches both qualitatively and quantitatively.
AB - Curvilinear Structure Segmentation (CSS) aims to predict the binary masks of curvilinear objects in photographs, such as blood vessels and road cracks. Despite impressive segmentation results, current CSS methods often struggle with disconnected segmentations and inaccurate edges, essentially caused by the distractions in the feature domain, which are from image domain-specific challenges, and by the imbalance of scales of thicker and thinner branches. To address these problems, we propose Distraction Mining and Dual Mutual Network (DMDMN) based on reverse attention and mutual learning to improve CSS. Specifically, DMDMN first extracts various levels of features from an input image using a plain network. Then, a Reverse Attention Module (RAM) is designed at each level to enhance the extracted features by identifying and removing false positive and negative distractions. Next is a Three-Head Fusion Module (THFM) at a separate level, which serves as an exchanger to mutually integrate the features from the branch segmentation head with those from the heads of skeleton extraction and edge detection. With the segmented results flowing back to RAMs, the previous two steps alternate several times till the final prediction of a high-quality segmentation, which is topologically connective along branches and pixel-wise accurate at edges. In addition, extensive experiments on five public datasets have demonstrated the superiority of the proposed DMDMN over state-of-the-art approaches both qualitatively and quantitatively.
KW - Curvilinear structure
KW - Multi-task learning
KW - Reverse attention
KW - Semantic segmentation
UR - http://www.scopus.com/inward/record.url?scp=105008771701&partnerID=8YFLogxK
U2 - 10.1007/s10489-025-06710-z
DO - 10.1007/s10489-025-06710-z
M3 - Article
AN - SCOPUS:105008771701
SN - 0924-669X
VL - 55
JO - Applied Intelligence
JF - Applied Intelligence
IS - 11
M1 - 817
ER -