Principal views selection based on growing graph convolution network for multi-view 3D model recognition

被引:0
作者
Liang, Qi [1 ]
Li, Qiang [1 ]
Nie, Weizhi [2 ]
Su, Yuting [2 ]
机构
[1] Tianjin Univ, Sch Microelect, Tianjin 300072, Peoples R China
[2] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-view; 3D model recognition; Growing graph; Deep neural network; NEURAL-NETWORK; OBJECT RETRIEVAL;
D O I
10.1007/s10489-022-03775-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the development of 3D technologies, 3D model recognition has attracted substantial attention in various areas, such as automatic driving, virtual/augmented reality, and computer-aided design. Many researchers are devoted to 3D model recognition and obtain some achievements in research. However, the abundant structure information of the 3D model also brings a huge challenge in model representation. In recent years, many researchers focus on classical computer vision technologies, which are utilized to represent the multi-view information of the 3D model. However, redundant visual information also brings a new challenge in model representation. In this paper, we focus on the multi-view 3D model data and propose a novel growing graph convolution network (GGCN) to handle the principal views selection problem, which can guarantee the performance of 3D model representation and effectively reduce the cost time. The proposed method mainly includes two modules: 1) principal views selection module: we utilize the selected views to describe the 3D model, which can effectively remove the redundant information and reduce computational complexity. 2) growing GCN module: we propose an effective growing GCN model, which focuses on gathering nodes that were less related to each other to ensure the result of multi-view fusion. It can indirectly retain the structure information and also reduce redundant information. In the process of graph growing, the GGCN model gradually adds view information to make up for the lack of characterization and guarantee the final performance. More specially, these two modules can guide each other to improve the performance of the principal views module and indirectly increase the final recognition accuracy. To evaluate the effectiveness of our proposed method, we test the classification accuracy and retrieval performance on the ModelNet40 dataset and ShapeNet dataset. The experimental results demonstrate the superiority of our proposed method.
引用
收藏
页码:5320 / 5336
页数:17
相关论文
共 59 条
[1]  
Allen M, 2008, 2008 INTERNATIONAL CONFERENCE ON INFORMATION PROCESSING IN SENSOR NETWORKS, PROCEEDINGS, P371, DOI 10.1109/IPSN.2008.45
[2]  
[Anonymous], 2017, EUR WORKSH 3 OBJ RET
[3]  
[Anonymous], 2015, P IEEE C COMPUTER VI, DOI [10.1109/CVPR.2015.7298801, DOI 10.1109/CVPR.2015.7298801]
[4]  
[Anonymous], 2012, Advances in Neural Information Processing Systems
[5]   GIFT: A Real-time and Scalable 3D Shape Search Engine [J].
Bai, Song ;
Bai, Xiang ;
Zhou, Zhichao ;
Zhang, Zhaoxiang ;
Latecki, Longin Jan .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5023-5032
[6]  
Bengio Yoshua, 2012, Neural Networks: Tricks of the Trade. Second Edition: LNCS 7700, P437, DOI 10.1007/978-3-642-35289-8_26
[7]   Voxel-based three-view hybrid parallel network for 3D object classification [J].
Cai, Weiwei ;
Liu, Dong ;
Ning, Xin ;
Wang, Chen ;
Xie, Guojie .
DISPLAYS, 2021, 69 (69)
[8]   VERAM: View-Enhanced Recurrent Attention Model for 3D Shape Classification [J].
Chen, Songle ;
Zheng, Lintao ;
Zhang, Yan ;
Sun, Zhixin ;
Xu, Kai .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2019, 25 (12) :3244-3257
[9]  
Dai GX, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P670
[10]  
Defferrard M, 2016, ADV NEUR IN, V29