ARM3D: Attention-based relation module for indoor 3D object detection

被引:14
|
作者
Lan, Yuqing [1 ]
Duan, Yao [1 ]
Liu, Chenyi [1 ]
Zhu, Chenyang [1 ]
Xiong, Yueshan [1 ]
Huang, Hui [2 ]
Xu, Kai [1 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Changsha 410073, Peoples R China
[2] Shenzhen Univ, Visual Comp Res Ctr, Shenzhen 518061, Peoples R China
基金
国家重点研发计划;
关键词
attention mechanism; scene understanding; relational reasoning; 3D indoor object detection; NETWORK;
D O I
10.1007/s41095-021-0252-6
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Relation contexts have been proved to be useful for many challenging vision tasks. In the field of 3D object detection, previous methods have been taking the advantage of context encoding, graph embedding, or explicit relation reasoning to extract relation contexts. However, there exist inevitably redundant relation contexts due to noisy or low-quality proposals. In fact, invalid relation contexts usually indicate underlying scene misunderstanding and ambiguity, which may, on the contrary, reduce the performance in complex scenes. Inspired by recent attention mechanism like Transformer, we propose a novel 3D attention-based relation module (ARM3D). It encompasses objectaware relation reasoning to extract pair-wise relation contexts among qualified proposals and an attention module to distribute attention weights towards different relation contexts. In this way, ARM3D can take full advantage of the useful relation contexts and filter those less relevant or even confusing contexts, which mitigates the ambiguity in detection. We have evaluated the effectiveness of ARM3D by plugging it into several state-of-the-art 3D object detectors and showing more accurate and robust detection results. Extensive experiments show the capability and generalization of ARM3D on 3D object detection. Our source code is available at https://github.com/lanlan96/ARM3D.
引用
收藏
页码:395 / 414
页数:20
相关论文
共 50 条
  • [1] ARM3D: Attention-based relation module for indoor 3D object detection
    Yuqing Lan
    Yao Duan
    Chenyi Liu
    Chenyang Zhu
    Yueshan Xiong
    Hui Huang
    Kai Xu
    Computational Visual Media, 2022, 8 : 395 - 414
  • [2] ARM3D: Attention-based relation module for indoor 3D object detection
    Yuqing Lan
    Yao Duan
    Chenyi Liu
    Chenyang Zhu
    Yueshan Xiong
    Hui Huang
    Kai Xu
    ComputationalVisualMedia, 2022, 8 (03) : 395 - 414
  • [3] Attention-based Proposals Refinement for 3D Object Detection
    Minh-Quan Dao
    Hery, Elwan
    Fremont, Vincent
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 197 - 205
  • [4] Hybrid Attention-Based 3D Object Detection with Differential Point Clouds
    Han, Guangjie
    Zhu, Yintian
    Liao, Lyuchao
    Yao, Huiwen
    Zhao, Zhaolin
    Zheng, Qi
    ELECTRONICS, 2022, 11 (23)
  • [5] Hierarchical Point Attention for Indoor 3D Object Detection
    Shu, Manli
    Xue, Le
    Yu, Ning
    Martin-Martin, Roberto
    Xiong, Calming
    Goldstein, Tom
    Niebles, Juan Carlos
    Xu, Ran
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 4245 - 4251
  • [6] Attention-Based Depth Distillation with 3D-Aware Positional Encoding for Monocular 3D Object Detection
    Wu, Zizhang
    Wu, Yunzhe
    Pu, Jian
    Li, Xianzhi
    Wang, Xiaoquan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 2892 - 2900
  • [7] Stereo 3D Object Detection Using a Feature Attention Module
    Zhao, Kexin
    Jiang, Rui
    He, Jun
    ALGORITHMS, 2023, 16 (12)
  • [8] Image attention transformer network for indoor 3D object detection
    REN KeYan
    YAN Tong
    HU ZhaoXin
    HAN HongGui
    ZHANG YunLu
    Science China(Technological Sciences), 2024, (07) : 2176 - 2190
  • [9] Image attention transformer network for indoor 3D object detection
    Ren, Keyan
    Yan, Tong
    Hu, Zhaoxin
    Han, Honggui
    Zhang, Yunlu
    SCIENCE CHINA-TECHNOLOGICAL SCIENCES, 2024, 67 (07) : 2176 - 2190
  • [10] Image attention transformer network for indoor 3D object detection
    REN KeYan
    YAN Tong
    HU ZhaoXin
    HAN HongGui
    ZHANG YunLu
    Science China(Technological Sciences), 2024, 67 (07) : 2176 - 2190