TY - GEN
T1 - CADNet
T2 - 2024 International Conference on Electronic Engineering and Information Systems, EEISS 2024
AU - Zhu, Canjie
AU - Sun, Huifang
AU - Lu, Mingfeng
AU - Zhang, Feng
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Monocular depth estimation is a crucial technology for comprehending scenes, and acquiring global contextual information is pivotal for enhancing depth estimation accuracy. Traditional approaches for incorporating global context information involve pooling feature maps of varying receptive field sizes. Nevertheless, they fail to address challenges such as object boundary distortion and the loss of local detail information caused by complex textures and geometric structures in scenes. To tackle these issues, this paper proposes a novel monocular depth estimation model called CADNet (Context-aggregated DCPPM monocular depth estimation network). This model leverages a multi-scale context aggregation module, DCPPM, to effectively aggregate local features into a global framework, thereby resolving the problem of local detail loss during network training. Experimental results demonstrate that the CADNet model surpasses the NewCRFs model in complex scene boundary detection and capturing local object details. Furthermore, with a 6.27% reduction in parameter count, the CADNet model achieves a noteworthy 9.82% decrease in Sq Rel error on the KITTI dataset and exhibits remarkable performance in general depth estimation metrics for both indoor and outdoor scenes.
AB - Monocular depth estimation is a crucial technology for comprehending scenes, and acquiring global contextual information is pivotal for enhancing depth estimation accuracy. Traditional approaches for incorporating global context information involve pooling feature maps of varying receptive field sizes. Nevertheless, they fail to address challenges such as object boundary distortion and the loss of local detail information caused by complex textures and geometric structures in scenes. To tackle these issues, this paper proposes a novel monocular depth estimation model called CADNet (Context-aggregated DCPPM monocular depth estimation network). This model leverages a multi-scale context aggregation module, DCPPM, to effectively aggregate local features into a global framework, thereby resolving the problem of local detail loss during network training. Experimental results demonstrate that the CADNet model surpasses the NewCRFs model in complex scene boundary detection and capturing local object details. Furthermore, with a 6.27% reduction in parameter count, the CADNet model achieves a noteworthy 9.82% decrease in Sq Rel error on the KITTI dataset and exhibits remarkable performance in general depth estimation metrics for both indoor and outdoor scenes.
KW - Contextual information aggregation
KW - Depth boundaries
KW - Fully connected conditional random field
KW - Monocular depth estimation
KW - Scene comprehension
UR - http://www.scopus.com/inward/record.url?scp=85201207630&partnerID=8YFLogxK
U2 - 10.1109/EEISS62553.2024.00035
DO - 10.1109/EEISS62553.2024.00035
M3 - Conference contribution
AN - SCOPUS:85201207630
T3 - Proceedings - 2024 International Conference on Electronic Engineering and Information Systems, EEISS 2024
SP - 161
EP - 165
BT - Proceedings - 2024 International Conference on Electronic Engineering and Information Systems, EEISS 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 13 January 2024 through 15 January 2024
ER -