TY - JOUR
T1 - Y-Net
T2 - Learning Domain Robust Feature Representation for ground camera image and large-scale image-based point cloud registration
AU - Liu, Weiquan
AU - Wang, Cheng
AU - Chen, Shuting
AU - Bian, Xuesheng
AU - Lai, Baiqi
AU - Shen, Xuelun
AU - Cheng, Ming
AU - Lai, Shang Hong
AU - Weng, Dongdong
AU - Li, Jonathan
N1 - Publisher Copyright:
© 2021 Elsevier Inc.
PY - 2021/12
Y1 - 2021/12
N2 - Registering the 2D images (2D space) with the 3D model of the environment (3D space) provides a promising solution to outdoor Augmented Reality (AR) virtual-real registration. In this work, we use the position and orientation of the ground camera image to synthesize a corresponding rendered image from the outdoor large-scale 3D image-based point cloud. To achieve the virtual-real registration, we indirectly establish the spatial relationship between 2D and 3D space by matching the above two kinds (2D/3D space) of cross-domain images. However, matching cross-domain images goes beyond the capability of handcrafted descriptors and existing deep neural networks. To address this issue, we propose an end-to-end network, Y-Net, to learn Domain Robust Feature Representations (DRFRs) for the cross-domain images. Besides, we introduce a cross-domain-constrained loss function that balances the loss in image content and cross-domain consistency of the feature representations. Experimental results show that the DRFRs simultaneously preserve the representation of image content and suppress the influence of independent domains. Furthermore, Y-Net outperforms the existing algorithms on extracting feature representations and achieves state-of-the-art performance in cross-domain image retrieval. Finally, we validate the Y-Net-based registration approach on campus to demonstrate its possible applicability.
AB - Registering the 2D images (2D space) with the 3D model of the environment (3D space) provides a promising solution to outdoor Augmented Reality (AR) virtual-real registration. In this work, we use the position and orientation of the ground camera image to synthesize a corresponding rendered image from the outdoor large-scale 3D image-based point cloud. To achieve the virtual-real registration, we indirectly establish the spatial relationship between 2D and 3D space by matching the above two kinds (2D/3D space) of cross-domain images. However, matching cross-domain images goes beyond the capability of handcrafted descriptors and existing deep neural networks. To address this issue, we propose an end-to-end network, Y-Net, to learn Domain Robust Feature Representations (DRFRs) for the cross-domain images. Besides, we introduce a cross-domain-constrained loss function that balances the loss in image content and cross-domain consistency of the feature representations. Experimental results show that the DRFRs simultaneously preserve the representation of image content and suppress the influence of independent domains. Furthermore, Y-Net outperforms the existing algorithms on extracting feature representations and achieves state-of-the-art performance in cross-domain image retrieval. Finally, we validate the Y-Net-based registration approach on campus to demonstrate its possible applicability.
KW - Cross-domain image
KW - Domain Robust Feature Representation (DRFR)
KW - Image patch matching
KW - Outdoor Augmented Reality
KW - Virtual-real registration
KW - Y-Net
UR - http://www.scopus.com/inward/record.url?scp=85117145614&partnerID=8YFLogxK
U2 - 10.1016/j.ins.2021.10.022
DO - 10.1016/j.ins.2021.10.022
M3 - Article
AN - SCOPUS:85117145614
SN - 0020-0255
VL - 581
SP - 655
EP - 677
JO - Information Sciences
JF - Information Sciences
ER -