PPF-Net: Efficient Multimodal 3D Object Detection with Pillar-Point Fusion

被引:0
|
作者
Zhang, Lingxiao [1 ]
Li, Changyong [1 ]
机构
[1] Xinjiang Univ, Coll Mech Engn, Urumqi 830017, Peoples R China
来源
ELECTRONICS | 2025年 / 14卷 / 04期
关键词
3D object detection; cross-modal data augmentation; sensor fusion; joint regression loss function;
D O I
10.3390/electronics14040685
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Detecting objects in 3D space using LiDAR is crucial for robotics and autonomous vehicles, but the sparsity of LiDAR-generated point clouds limits performance. Camera images, rich in semantic information, can effectively compensate for this limitation. We propose a simpler yet effective multimodal fusion framework to enhance 3D object detection without complex network designs. We introduce a cross-modal GT-Paste data augmentation method to address challenges like 2D object acquisition and occlusions from added objects. To better integrate image features with sparse point clouds, we propose Pillar-Point Fusion (PPF), which projects non-empty pillars onto image feature maps and uses an attention mechanism to map semantic features from pillars to their constituent points, fusing them with the points' geometric features. Additionally, we design the BD-IoU loss function, which measures 3D bounding box similarity, and a joint regression loss combining BD-IoU and Smooth L1, effectively guiding model training. Our framework achieves consistent improvements across KITTI benchmarks. On the validation set, PFF (PV-RCNN) achieves at least 1.84% AP improvement in Cyclist detection performance across all difficulty levels compared to other multimodal SOTA methods. On the test set, PPF-Net excels in pedestrian detection for moderate and hard difficulty levels and achieves the best results in low-beam LiDAR scenarios.
引用
收藏
页数:21
相关论文
共 50 条
  • [31] TinyPillarNet: Tiny Pillar-Based Network for 3D Point Cloud Object Detection at Edge
    Li, Yishi
    Zhang, Yuhao
    Lai, Rui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (03) : 1772 - 1785
  • [32] EPAWFusion: multimodal fusion for 3D object detection based on enhanced points and adaptive weights
    Sun, Xiang
    Song, Shaojing
    Wu, Fan
    Lu, Tingting
    Li, Bohao
    Miao, Zhiqing
    JOURNAL OF APPLIED REMOTE SENSING, 2024, 18 (01)
  • [33] SAMFusion: Sensor-Adaptive Multimodal Fusion for 3D Object Detection in Adverse Weather
    Palladin, Edoardo
    Dietze, Roland
    Narayanan, Praveen
    Bijelic, Mario
    Heide, Felix
    COMPUTER VISION - ECCV 2024, PT LXI, 2025, 15119 : 484 - 503
  • [34] Transformer-Based Optimized Multimodal Fusion for 3D Object Detection in Autonomous Driving
    Alaba, Simegnew Yihunie
    Ball, John E.
    IEEE ACCESS, 2024, 12 : 50165 - 50176
  • [35] Point cloud 3D object detection algorithm based on local information fusion
    Zhang, Linjie
    Chai, Zhilei
    Wang, Ning
    Zhejiang Daxue Xuebao (Gongxue Ban)/Journal of Zhejiang University (Engineering Science), 2024, 58 (11): : 2219 - 2229
  • [36] Real Pseudo-Lidar Point Cloud Fusion for 3D Object Detection
    Fan, Xiangsuo
    Xiao, Dachuan
    Cai, Dengsheng
    Ding, Wentao
    ELECTRONICS, 2023, 12 (18)
  • [37] 3D object detection based on fusion of point cloud and image by mutual attention
    Chen J.-Y.
    Bai T.-Y.
    Zhao L.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2021, 29 (09): : 2247 - 2254
  • [38] CenterTransFuser: radar point cloud and visual information fusion for 3D object detection
    Yan Li
    Kai Zeng
    Tao Shen
    EURASIP Journal on Advances in Signal Processing, 2023
  • [39] STFNET: Sparse Temporal Fusion for 3D Object Detection in LiDAR Point Cloud
    Meng, Xin
    Zhou, Yuan
    Ma, Jun
    Jiang, Fangdi
    Qi, Yongze
    Wang, Cui
    Kim, Jonghyuk
    Wang, Shifeng
    IEEE SENSORS JOURNAL, 2025, 25 (03) : 5866 - 5877
  • [40] Research on 3D Object Detection Based on Laser Point Cloud and Image Fusion
    Liu Y.
    Yu F.
    Zhang X.
    Chen Z.
    Qin D.
    Jixie Gongcheng Xuebao/Journal of Mechanical Engineering, 2022, 58 (24): : 289 - 299