TY - JOUR
T1 - Multi-task information propagation network for camouflaged object detection
AU - Zhao, Fei
AU - Lou, Wenzhong
AU - Feng, Hengzhen
AU - Ding, Nanxi
AU - Li, Chenglong
AU - Ma, Wenlong
N1 - Publisher Copyright:
© 2025 Elsevier Inc.
PY - 2025/7
Y1 - 2025/7
N2 - Camouflaged object detection (COD) aims to discover camouflaged objects embedded in the background. However, most existing COD methods focus solely on extracting single-task features for inference, and rely on relatively simple supervision data, which hampers their generalization ability in real-world deployment. To meet these challenges, we design a novel camouflaged object detection framework, the Multi-Task Information Propagation Network (MIP-Net). In the feature extraction stage, the Multi-Task Feature Extraction Module (MFEM) learns multi-task feature representations that preserve boundary, texture, and object information to enhance COD performance. During the feature fusion encoding stage, the Cross-Level Multi-Task Fusion Propagation (CMFP) module effectively integrates multi-task features from different levels, adaptively weighting the encoder's features so that the decoder receives more effective information for improved COD outcomes. In the feature decoding inference stage, the Cross-Layer Progressive Refinement Module (CSRM) captures the complementarity of cross-level features, further refining the final inference results. We conducted comprehensive experiments on four benchmark datasets: CHAMELEON, CAMO, COD10K, and NC4K. The comprehensive results prove that the proposed MIP-Net framework outperforms existing methods, showcasing its effectiveness and robustness.
AB - Camouflaged object detection (COD) aims to discover camouflaged objects embedded in the background. However, most existing COD methods focus solely on extracting single-task features for inference, and rely on relatively simple supervision data, which hampers their generalization ability in real-world deployment. To meet these challenges, we design a novel camouflaged object detection framework, the Multi-Task Information Propagation Network (MIP-Net). In the feature extraction stage, the Multi-Task Feature Extraction Module (MFEM) learns multi-task feature representations that preserve boundary, texture, and object information to enhance COD performance. During the feature fusion encoding stage, the Cross-Level Multi-Task Fusion Propagation (CMFP) module effectively integrates multi-task features from different levels, adaptively weighting the encoder's features so that the decoder receives more effective information for improved COD outcomes. In the feature decoding inference stage, the Cross-Layer Progressive Refinement Module (CSRM) captures the complementarity of cross-level features, further refining the final inference results. We conducted comprehensive experiments on four benchmark datasets: CHAMELEON, CAMO, COD10K, and NC4K. The comprehensive results prove that the proposed MIP-Net framework outperforms existing methods, showcasing its effectiveness and robustness.
KW - Camouflaged object detection
KW - Feature fusion
KW - Feature propagation
KW - Multi-task learning
KW - Swin transformer
UR - http://www.scopus.com/inward/record.url?scp=105001364823&partnerID=8YFLogxK
U2 - 10.1016/j.dsp.2025.105172
DO - 10.1016/j.dsp.2025.105172
M3 - Article
AN - SCOPUS:105001364823
SN - 1051-2004
VL - 162
JO - Digital Signal Processing: A Review Journal
JF - Digital Signal Processing: A Review Journal
M1 - 105172
ER -