TY - GEN
T1 - 3D-TRANS
T2 - 10th International Conference on Automation, Robotics, and Applications, ICARA 2024
AU - Huang, Hao
AU - Yuan, Shuaihang
AU - Wen, Congcong
AU - Hao, Yu
AU - Fang, Yi
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Motivated by the intuition that one can find correspondences between two 3D shapes in a coarse-to-fine manner, we propose 3D-TRANS, a novel learning-based model to match deformable 3D shapes hierarchically. Specifically, we first propose a new subspace feature learning approach that takes advantage of local and global geometric structures of 3D shapes. Then, we design a new hierarchical 3D transformer that adaptively learns the spatial relationships of different space partitions, which correspond to different regions of shapes at multiple scales. Inside the transformer, low-rank self-attention is utilized to extract discriminative region-aware shape descriptors, capturing both short- and long-range geometric dependencies of different shape regions at an economical computational cost. Finally, we develop a new multiple templates fusion scheme to fuse multiple deformed templates using region-aware shape descriptors to predict correspondences between input shapes and templates. We demonstrate that our approach improves over the state-of-the-art results on the difficult FAUST-inter and -intra shapes. Our model also performs well on real partial shapes from the SCAPE dataset and non-human shapes from the TOSCA dataset.
AB - Motivated by the intuition that one can find correspondences between two 3D shapes in a coarse-to-fine manner, we propose 3D-TRANS, a novel learning-based model to match deformable 3D shapes hierarchically. Specifically, we first propose a new subspace feature learning approach that takes advantage of local and global geometric structures of 3D shapes. Then, we design a new hierarchical 3D transformer that adaptively learns the spatial relationships of different space partitions, which correspond to different regions of shapes at multiple scales. Inside the transformer, low-rank self-attention is utilized to extract discriminative region-aware shape descriptors, capturing both short- and long-range geometric dependencies of different shape regions at an economical computational cost. Finally, we develop a new multiple templates fusion scheme to fuse multiple deformed templates using region-aware shape descriptors to predict correspondences between input shapes and templates. We demonstrate that our approach improves over the state-of-the-art results on the difficult FAUST-inter and -intra shapes. Our model also performs well on real partial shapes from the SCAPE dataset and non-human shapes from the TOSCA dataset.
KW - K-d tree
KW - Shape registration
KW - Transformer
KW - hierarchical shape analysis
UR - http://www.scopus.com/inward/record.url?scp=85197356994&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85197356994&partnerID=8YFLogxK
U2 - 10.1109/ICARA60736.2024.10552931
DO - 10.1109/ICARA60736.2024.10552931
M3 - Conference contribution
AN - SCOPUS:85197356994
T3 - 2024 10th International Conference on Automation, Robotics, and Applications, ICARA 2024
SP - 536
EP - 540
BT - 2024 10th International Conference on Automation, Robotics, and Applications, ICARA 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 22 February 2024 through 24 February 2024
ER -