TY - JOUR
T1 - CompleteDT
T2 - Point cloud completion with information-perception transformers
AU - Li, Jun
AU - Guo, Shangwei
AU - Wang, Luhan
AU - Han, Shaokun
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2024/8/1
Y1 - 2024/8/1
N2 - In this work, we propose a novel point cloud completion network, called CompleteDT. To fully capture the 3D geometric structure of point clouds, we introduce an Information-Perception Transformer (IPT) that can simultaneously capture local features and global geometric relations. CompleteDT comprises a Feature Encoder, Query Generator, and Query Decoder. Feature Encoder extracts local features from multi-resolution point clouds to capture intricate geometrical structures. Query Generator uses the proposed IPT, utilizing the Point Local Attention (PLA) and Point Global Attention (PGA) modules, to learn local features and global correlations, and generate query features that represent predicted point clouds. The PLA captures local information within local points by adaptively measuring weights of neighboring points, while PGA adapts multi-head self-attention by transforming it into a layer-by-layer form where each head learns global features in a high-dimensional space of different dimensions. By dense connections, the module allows for direct information exchange between each head and facilitates the capture of long global correlations. By combining the strengths of both PLA and PGA, the IPT can fully leverage local and global features to facilitate CompleteDT to complete point clouds. Lastly, the query features undergo refining to generate a complete point cloud through the Query Decoder. Our experimental results demonstrate that CompleteDT outperforms current state-of-the-art methods, effectively learning from incomplete inputs and predicting complete outputs.
AB - In this work, we propose a novel point cloud completion network, called CompleteDT. To fully capture the 3D geometric structure of point clouds, we introduce an Information-Perception Transformer (IPT) that can simultaneously capture local features and global geometric relations. CompleteDT comprises a Feature Encoder, Query Generator, and Query Decoder. Feature Encoder extracts local features from multi-resolution point clouds to capture intricate geometrical structures. Query Generator uses the proposed IPT, utilizing the Point Local Attention (PLA) and Point Global Attention (PGA) modules, to learn local features and global correlations, and generate query features that represent predicted point clouds. The PLA captures local information within local points by adaptively measuring weights of neighboring points, while PGA adapts multi-head self-attention by transforming it into a layer-by-layer form where each head learns global features in a high-dimensional space of different dimensions. By dense connections, the module allows for direct information exchange between each head and facilitates the capture of long global correlations. By combining the strengths of both PLA and PGA, the IPT can fully leverage local and global features to facilitate CompleteDT to complete point clouds. Lastly, the query features undergo refining to generate a complete point cloud through the Query Decoder. Our experimental results demonstrate that CompleteDT outperforms current state-of-the-art methods, effectively learning from incomplete inputs and predicting complete outputs.
KW - 3D point cloud
KW - 3D reconstruction
KW - Point cloud completion
KW - Transformer
UR - http://www.scopus.com/inward/record.url?scp=85192301823&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2024.127790
DO - 10.1016/j.neucom.2024.127790
M3 - Article
AN - SCOPUS:85192301823
SN - 0925-2312
VL - 592
JO - Neurocomputing
JF - Neurocomputing
M1 - 127790
ER -