TY - GEN
T1 - Splatting-Based View Synthesis for Self-supervised Monocular Depth Estimation
AU - Liu, Jiahao
AU - Leng, Jianghao
AU - Liu, Bo
AU - Huang, Wenyi
AU - Sun, Chao
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Self-supervised method has shown great potential in monocular depth estimation, since it does not need expensive ground-truth depth labels but only uses the photometric error of synthesized images as the supervision signal. However, although many methods have been proposed to improve its performance, the occlusion problem has not been clearly handled. This paper introduces a novel view synthesis module to deal with occluded pixels in the process of image reconstruction. Specifically, we use bilinear splatting to forward warp the source image, and average pixels projected to the same location by the predicted depth. In addition, a valid pixel mask is generated with projection to ignore invalid pixels. The proposed approach clearly handles overlapping pixels and invalid areas of the synthesized image, thus improving the performance of self-supervised learning. We conduct various experiments, and the results show that our model can generate clear and complete depth maps and achieves state-of-the-art performance.
AB - Self-supervised method has shown great potential in monocular depth estimation, since it does not need expensive ground-truth depth labels but only uses the photometric error of synthesized images as the supervision signal. However, although many methods have been proposed to improve its performance, the occlusion problem has not been clearly handled. This paper introduces a novel view synthesis module to deal with occluded pixels in the process of image reconstruction. Specifically, we use bilinear splatting to forward warp the source image, and average pixels projected to the same location by the predicted depth. In addition, a valid pixel mask is generated with projection to ignore invalid pixels. The proposed approach clearly handles overlapping pixels and invalid areas of the synthesized image, thus improving the performance of self-supervised learning. We conduct various experiments, and the results show that our model can generate clear and complete depth maps and achieves state-of-the-art performance.
KW - KITTI dataset
KW - monocular depth estimation
KW - self-supervised learning
KW - view synthesis
UR - http://www.scopus.com/inward/record.url?scp=85164247379&partnerID=8YFLogxK
U2 - 10.1109/EECR56827.2023.10150161
DO - 10.1109/EECR56827.2023.10150161
M3 - Conference contribution
AN - SCOPUS:85164247379
T3 - 2023 9th International Conference on Electrical Engineering, Control and Robotics, EECR 2023
SP - 274
EP - 279
BT - 2023 9th International Conference on Electrical Engineering, Control and Robotics, EECR 2023
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 9th International Conference on Electrical Engineering, Control and Robotics, EECR 2023
Y2 - 24 February 2023 through 26 February 2023
ER -