Laplacian Mesh Transformer: Dual Attention and Topology Aware Network for 3D Mesh Classification and Segmentation

被引:13
作者
Li, Xiao-Juan [1 ,2 ]
Yang, Jie [1 ,2 ]
Zhang, Fang-Lue [3 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Victoria Univ Wellington, Wellington, New Zealand
来源
COMPUTER VISION, ECCV 2022, PT XXIX | 2022年 / 13689卷
基金
中国国家自然科学基金;
关键词
Laplacian EigenVector; Transformer; Attention mechanism; Topology aware; Shape segmentation & classification; SHAPE SEGMENTATION;
D O I
10.1007/978-3-031-19818-2_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning-based approaches for shape understanding and processing tasks have attracted considerable attention. Despite the great progress that has been made, the existing approaches fail to efficiently capture sophisticated structure information and critical part features simultaneously, limiting their capability of providing discriminative deep shape features. To address the above issue, we proposed a novel deep learning framework, Laplacian Mesh Transformer, to extract the critical structure and geometry features. We introduce a dual attention mechanism, where the 1(st) level self-attention mechanism is used to capture the critical partial/local structure and geometric information on the entire mesh, and the 2(nd) level is to fuse the geometrical and structural features together with the learned importance according to a specific downstream task. More particularly, Laplacian spectral decomposition is adopted as our basic structure representation given its ability to describe shape topology (connectivity of triangles). Our approach builds a hierarchical structure to process shape features from fine to coarse using the dual attention mechanism, which is stable under the isometric transformations. It enables an effective feature extraction that can tackle 3D meshes with complex structure and geometry efficiently in various shape analysis tasks, such as shape segmentation and classification. Extensive experiments on the standard benchmarks show that our method outperforms state-of-the-art methods.
引用
收藏
页码:541 / 560
页数:20
相关论文
共 89 条
[1]  
Achlioptas P., 2018, Learning representations and generative models for 3d point clouds. pages, P40
[2]  
Ahmed E, 2019, Arxiv, DOI arXiv:1808.01462
[3]  
[Anonymous], 2017, INT C LEARNING REPRE
[4]   3D Semantic Parsing of Large-Scale Indoor Spaces [J].
Armeni, Iro ;
Sener, Ozan ;
Zamir, Amir R. ;
Jiang, Helen ;
Brilakis, Ioannis ;
Fischer, Martin ;
Savarese, Silvio .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1534-1543
[5]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[6]  
Boscaini D, 2016, ADV NEUR IN, V29
[7]  
Boski M, 2017, 2017 10TH INTERNATIONAL WORKSHOP ON MULTIDIMENSIONAL (ND) SYSTEMS (NDS)
[8]  
Botsch Mario, 2010, Polygon Mesh Processing
[9]   Geometric Deep Learning Going beyond Euclidean data [J].
Bronstein, Michael M. ;
Bruna, Joan ;
LeCun, Yann ;
Szlam, Arthur ;
Vandergheynst, Pierre .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (04) :18-42
[10]  
Carion N, 2020, Arxiv, DOI arXiv:2005.12872