MFF-PR: Point Cloud and Image Multi-modal Feature Fusion for Place Recognition

被引:5
|
作者
Liu, Wenlei [1 ]
Fei, Jiajun [1 ]
Zhu, Ziyu [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci, Beijing, Peoples R China
来源
2022 IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY (ISMAR 2022) | 2022年
关键词
Place recognition; Multi-modal fusion; SLAM; Point cloud and image;
D O I
10.1109/ISMAR55827.2022.00082
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Place recognition technology can eliminate cumulative errors and thus plays a vital role in autonomous driving. In this paper, the composite feature of point cloud and image data is obtained by multi-modal feature fusion, thereby improving positioning accuracy. Semantic features, instance features, topological features, and image texture features are integrated to obtain comprehensive features, presenting strong robustness and complex scene expression abilities. Topological features consist of intra-class features and inter-instance features, which allow users to obtain more comprehensive scene structure information. The place recognition methods of data-level fusion and feature-level fusion based on point cloud and image are compared. This paper verifies the proposed method on SemanticKitti and nuScenes datasets. The results show that it outperforms state-of-the-art place recognition methods.
引用
收藏
页码:647 / 655
页数:9
相关论文
共 50 条
  • [1] PRFusion: Toward Effective and Robust Multi-Modal Place Recognition With Image and Point Cloud Fusion
    Wang, Sijie
    Kang, Qiyu
    She, Rui
    Zhao, Kai
    Song, Yang
    Tay, Wee Peng
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, : 20523 - 20534
  • [2] Electromagnetic signal feature fusion and recognition based on multi-modal deep learning
    Hou C.
    Zhang X.
    Chen X.
    International Journal of Performability Engineering, 2020, 16 (06): : 941 - 949
  • [3] On Multi-modal Fusion for Freehand Gesture Recognition
    Schak, Monika
    Gepperth, Alexander
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 : 862 - 873
  • [4] Low-level fusion of audio and video feature for multi-modal emotion recognition
    Wimmer, Matthias
    Schuller, Bjoern
    Arsic, Dejan
    Rigoll, Gerhard
    Radig, Bernd
    VISAPP 2008: PROCEEDINGS OF THE THIRD INTERNATIONAL CONFERENCE ON COMPUTER VISION THEORY AND APPLICATIONS, VOL 2, 2008, : 145 - +
  • [5] Gesture recognition based on multi-modal feature weight
    Duan, Haojie
    Sun, Ying
    Cheng, Wentao
    Jiang, Du
    Yun, Juntong
    Liu, Ying
    Liu, Yibo
    Zhou, Dalin
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (05)
  • [6] Recognition of multi-modal fusion images with irregular interference
    Wang, Yawei
    Chen, Yifei
    Wang, Dongfeng
    PEERJ COMPUTER SCIENCE, 2022, 8
  • [7] ATTENTION DRIVEN FUSION FOR MULTI-MODAL EMOTION RECOGNITION
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3227 - 3231
  • [8] Human activity recognition based on multi-modal fusion
    Cheng Zhang
    Tianqi Zu
    Yibin Hou
    Jian He
    Shengqi Yang
    Ruihai Dong
    CCF Transactions on Pervasive Computing and Interaction, 2023, 5 : 321 - 332
  • [9] Human activity recognition based on multi-modal fusion
    Zhang, Cheng
    Zu, Tianqi
    Hou, Yibin
    He, Jian
    Yang, Shengqi
    Dong, Ruihai
    CCF TRANSACTIONS ON PERVASIVE COMPUTING AND INTERACTION, 2023, 5 (03) : 321 - 332
  • [10] Heterogeneous Feature Fusion Approach for Multi-Modal Indoor Localization
    Zhou, Junyi
    Huang, Kaixuan
    Tang, Siyu
    Zhang, Shunqing
    2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024, 2024,