CWGA-Net: Center-Weighted Graph Attention Network for 3D object detection from point clouds

被引:0
作者
Shu, Jun [1 ,2 ]
Wu, Qi [1 ]
Tan, Liang [1 ]
Shu, Xinyi [3 ]
Wan, Fengchun [4 ]
机构
[1] Hubei Univ Technol, Sch Elect & Elect Engn, Nanli Rd 28, Wuhan 430068, Peoples R China
[2] Hubei Univ Technol, Hubei Key Lab High Efficiency Utilizat Solar Energ, Nanli Rd 28, Wuhan 430068, Peoples R China
[3] Univ Melbourne, Fac Sci, Melbourne, Vic 3010, Australia
[4] Hunan Xianbu Informat Co Ltd, Changsha 410116, Peoples R China
关键词
Autonomous driving; 3D object detection; Local graph encoding; Center-weighted cross-attention; Cross-feature fusion module;
D O I
10.1016/j.imavis.2024.105314
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The precision of 3D object detection from unevenly distributed outdoor point clouds is critical in autonomous driving perception systems. Current point-based detectors employ self-attention and graph convolution to establish contextual relationships between point clouds; however, they often introduce weakly correlated redundant information, leading to blurred geometric details and false detections. To address this issue, a novel Center-weighted Graph Attention Network (CWGA-Net) has been proposed to fuse geometric and semantic similarities for weighting cross-attention scores, thereby capturing precise fine-grained geometric features. CWGA-Net initially constructs and encodes local graphs between foreground points, establishing connections between point clouds from geometric and semantic dimensions. Subsequently, center-weighted cross-attention is utilized to compute the contextual relationships between vertices within the graph, and geometric and semantic similarities between vertices are fused to weight attention scores, thereby extracting strongly related geometric shape features. Finally, a cross-feature fusion Module is introduced to deeply fuse high and low- resolution features to compensate for the information loss during downsampling. Experiments conducted on the KITTI and Waymo datasets demonstrate that the network achieves superior detection capabilities, outperforming state-of-the-art point-based single-stage methods in terms of average precision metrics while maintaining good speed.
引用
收藏
页数:12
相关论文
共 48 条
  • [1] A Survey on 3D Object Detection Methods for Autonomous Driving Applications
    Arnold, Eduardo
    Al-Jarrah, Omar Y.
    Dianati, Mehrdad
    Fallah, Saber
    Oxtoby, David
    Mouzakitis, Alex
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (10) : 3782 - 3795
  • [2] SA-Det3D: Self-Attention Based Context-Aware 3D Object Detection
    Bhattacharyya, Prarthana
    Huang, Chengjie
    Czarnecki, Krzysztof
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 3022 - 3031
  • [3] Chen C, 2022, AAAI CONF ARTIF INTE, P221
  • [4] A Hierarchical Graph Network for 3D Object Detection on Point Clouds
    Chen, Jintai
    Lei, Biwen
    Song, Qingyu
    Ying, Haochao
    Chen, Danny Z.
    Wu, Jian
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 389 - 398
  • [5] Deng JJ, 2021, AAAI CONF ARTIF INTE, V35, P1201
  • [6] Dong S., 2022, Adv. Neural Inf. Process. Syst., V35, P11615
  • [7] Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
  • [8] Voxel Set Transformer: A Set-to-Set Approach to 3D Object Detection from Point Clouds
    He, Chenhang
    Li, Ruihuang
    Li, Shuai
    Zhang, Lei
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 8407 - 8417
  • [9] He CH, 2020, PROC CVPR IEEE, P11870, DOI 10.1109/CVPR42600.2020.01189
  • [10] He QD, 2022, AAAI CONF ARTIF INTE, P870